Dec 10 14:24:09 crc systemd[1]: Starting Kubernetes Kubelet... Dec 10 14:24:09 crc restorecon[4684]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:09 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 14:24:10 crc restorecon[4684]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 10 14:24:10 crc kubenswrapper[4847]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.594190 4847 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598424 4847 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598449 4847 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598455 4847 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598462 4847 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598469 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598474 4847 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598481 4847 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598486 4847 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598491 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598497 4847 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598503 4847 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598510 4847 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598516 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598523 4847 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598530 4847 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598538 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598544 4847 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598549 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598555 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598560 4847 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598566 4847 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598571 4847 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598576 4847 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598582 4847 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598588 4847 feature_gate.go:330] unrecognized feature gate: Example Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598593 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598598 4847 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598604 4847 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598608 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598614 4847 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598619 4847 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598624 4847 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598629 4847 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598634 4847 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598639 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598645 4847 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598650 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598656 4847 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598661 4847 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598666 4847 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598681 4847 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598688 4847 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598693 4847 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598699 4847 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598704 4847 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598732 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598738 4847 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598743 4847 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598748 4847 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598753 4847 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598758 4847 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598763 4847 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598768 4847 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598774 4847 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598779 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598784 4847 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598791 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598798 4847 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598805 4847 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598811 4847 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598816 4847 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598822 4847 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598827 4847 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598832 4847 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598838 4847 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598843 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598848 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598853 4847 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598858 4847 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598863 4847 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.598869 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599626 4847 flags.go:64] FLAG: --address="0.0.0.0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599647 4847 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599660 4847 flags.go:64] FLAG: --anonymous-auth="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599668 4847 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599676 4847 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599683 4847 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599692 4847 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599700 4847 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599728 4847 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599735 4847 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599741 4847 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599748 4847 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599754 4847 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599760 4847 flags.go:64] FLAG: --cgroup-root="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599766 4847 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599772 4847 flags.go:64] FLAG: --client-ca-file="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599778 4847 flags.go:64] FLAG: --cloud-config="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599784 4847 flags.go:64] FLAG: --cloud-provider="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599789 4847 flags.go:64] FLAG: --cluster-dns="[]" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599796 4847 flags.go:64] FLAG: --cluster-domain="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599802 4847 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599809 4847 flags.go:64] FLAG: --config-dir="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599815 4847 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599821 4847 flags.go:64] FLAG: --container-log-max-files="5" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599829 4847 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599835 4847 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599841 4847 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599849 4847 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599855 4847 flags.go:64] FLAG: --contention-profiling="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599861 4847 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599868 4847 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599875 4847 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599881 4847 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599889 4847 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599895 4847 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599902 4847 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599908 4847 flags.go:64] FLAG: --enable-load-reader="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599914 4847 flags.go:64] FLAG: --enable-server="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599922 4847 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599930 4847 flags.go:64] FLAG: --event-burst="100" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599936 4847 flags.go:64] FLAG: --event-qps="50" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599942 4847 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599948 4847 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599954 4847 flags.go:64] FLAG: --eviction-hard="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599961 4847 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599967 4847 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599973 4847 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599979 4847 flags.go:64] FLAG: --eviction-soft="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599985 4847 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599991 4847 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.599997 4847 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600003 4847 flags.go:64] FLAG: --experimental-mounter-path="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600009 4847 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600015 4847 flags.go:64] FLAG: --fail-swap-on="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600021 4847 flags.go:64] FLAG: --feature-gates="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600028 4847 flags.go:64] FLAG: --file-check-frequency="20s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600034 4847 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600040 4847 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600047 4847 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600053 4847 flags.go:64] FLAG: --healthz-port="10248" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600059 4847 flags.go:64] FLAG: --help="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600065 4847 flags.go:64] FLAG: --hostname-override="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600072 4847 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600078 4847 flags.go:64] FLAG: --http-check-frequency="20s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600084 4847 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600090 4847 flags.go:64] FLAG: --image-credential-provider-config="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600096 4847 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600102 4847 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600108 4847 flags.go:64] FLAG: --image-service-endpoint="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600114 4847 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600121 4847 flags.go:64] FLAG: --kube-api-burst="100" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600128 4847 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600135 4847 flags.go:64] FLAG: --kube-api-qps="50" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600141 4847 flags.go:64] FLAG: --kube-reserved="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600147 4847 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600153 4847 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600159 4847 flags.go:64] FLAG: --kubelet-cgroups="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600166 4847 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600172 4847 flags.go:64] FLAG: --lock-file="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600178 4847 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600184 4847 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600190 4847 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600199 4847 flags.go:64] FLAG: --log-json-split-stream="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600205 4847 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600211 4847 flags.go:64] FLAG: --log-text-split-stream="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600216 4847 flags.go:64] FLAG: --logging-format="text" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600223 4847 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600229 4847 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600235 4847 flags.go:64] FLAG: --manifest-url="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600241 4847 flags.go:64] FLAG: --manifest-url-header="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600249 4847 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600255 4847 flags.go:64] FLAG: --max-open-files="1000000" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600263 4847 flags.go:64] FLAG: --max-pods="110" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600269 4847 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600275 4847 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600281 4847 flags.go:64] FLAG: --memory-manager-policy="None" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600287 4847 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600293 4847 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600299 4847 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600305 4847 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600319 4847 flags.go:64] FLAG: --node-status-max-images="50" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600325 4847 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600332 4847 flags.go:64] FLAG: --oom-score-adj="-999" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600338 4847 flags.go:64] FLAG: --pod-cidr="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600344 4847 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600352 4847 flags.go:64] FLAG: --pod-manifest-path="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600358 4847 flags.go:64] FLAG: --pod-max-pids="-1" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600366 4847 flags.go:64] FLAG: --pods-per-core="0" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600372 4847 flags.go:64] FLAG: --port="10250" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600378 4847 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600385 4847 flags.go:64] FLAG: --provider-id="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600391 4847 flags.go:64] FLAG: --qos-reserved="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600397 4847 flags.go:64] FLAG: --read-only-port="10255" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600404 4847 flags.go:64] FLAG: --register-node="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600410 4847 flags.go:64] FLAG: --register-schedulable="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600416 4847 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600430 4847 flags.go:64] FLAG: --registry-burst="10" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600437 4847 flags.go:64] FLAG: --registry-qps="5" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600443 4847 flags.go:64] FLAG: --reserved-cpus="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600448 4847 flags.go:64] FLAG: --reserved-memory="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600456 4847 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600462 4847 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600468 4847 flags.go:64] FLAG: --rotate-certificates="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600474 4847 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600480 4847 flags.go:64] FLAG: --runonce="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600487 4847 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600495 4847 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600503 4847 flags.go:64] FLAG: --seccomp-default="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600510 4847 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600518 4847 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600525 4847 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600533 4847 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600540 4847 flags.go:64] FLAG: --storage-driver-password="root" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600549 4847 flags.go:64] FLAG: --storage-driver-secure="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600556 4847 flags.go:64] FLAG: --storage-driver-table="stats" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600564 4847 flags.go:64] FLAG: --storage-driver-user="root" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600572 4847 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600579 4847 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600588 4847 flags.go:64] FLAG: --system-cgroups="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600596 4847 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600607 4847 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600615 4847 flags.go:64] FLAG: --tls-cert-file="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600622 4847 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600632 4847 flags.go:64] FLAG: --tls-min-version="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600639 4847 flags.go:64] FLAG: --tls-private-key-file="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600646 4847 flags.go:64] FLAG: --topology-manager-policy="none" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600653 4847 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600661 4847 flags.go:64] FLAG: --topology-manager-scope="container" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600669 4847 flags.go:64] FLAG: --v="2" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600680 4847 flags.go:64] FLAG: --version="false" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600690 4847 flags.go:64] FLAG: --vmodule="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600698 4847 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.600706 4847 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600882 4847 feature_gate.go:330] unrecognized feature gate: Example Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600891 4847 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600896 4847 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600902 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600907 4847 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600912 4847 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600917 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600923 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600928 4847 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600933 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600940 4847 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600945 4847 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600953 4847 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600960 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600969 4847 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600974 4847 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600980 4847 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600987 4847 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600993 4847 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.600999 4847 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601005 4847 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601010 4847 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601015 4847 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601020 4847 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601025 4847 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601030 4847 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601036 4847 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601042 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601047 4847 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601052 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601057 4847 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601063 4847 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601068 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601073 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601078 4847 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601083 4847 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601089 4847 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601094 4847 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601099 4847 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601104 4847 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601109 4847 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601115 4847 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601120 4847 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601126 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601131 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601136 4847 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601144 4847 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601149 4847 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601155 4847 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601160 4847 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601165 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601170 4847 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601175 4847 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601180 4847 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601185 4847 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601190 4847 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601195 4847 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601200 4847 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601205 4847 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601211 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601216 4847 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601222 4847 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601229 4847 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601235 4847 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601241 4847 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601246 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601252 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601257 4847 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601262 4847 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601267 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.601274 4847 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.601459 4847 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.611915 4847 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.612017 4847 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612551 4847 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612588 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612596 4847 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612603 4847 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612610 4847 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612622 4847 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612632 4847 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612639 4847 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612647 4847 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612655 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612670 4847 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612677 4847 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612683 4847 feature_gate.go:330] unrecognized feature gate: Example Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612690 4847 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612696 4847 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612702 4847 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612729 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612736 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612742 4847 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612748 4847 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612755 4847 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612761 4847 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612773 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612780 4847 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612786 4847 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612792 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612798 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612804 4847 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612810 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612816 4847 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612822 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612828 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612836 4847 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612842 4847 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612847 4847 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612859 4847 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612868 4847 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612875 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612882 4847 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612890 4847 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612898 4847 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612905 4847 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612912 4847 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612919 4847 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612925 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612932 4847 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612939 4847 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612952 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612959 4847 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612965 4847 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612972 4847 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612978 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612987 4847 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.612995 4847 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613003 4847 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613011 4847 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613017 4847 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613023 4847 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613030 4847 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613036 4847 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613047 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613053 4847 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613061 4847 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613068 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613077 4847 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613084 4847 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613090 4847 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613096 4847 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613103 4847 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613108 4847 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613114 4847 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.613125 4847 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613963 4847 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613985 4847 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.613993 4847 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614000 4847 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614008 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614017 4847 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614023 4847 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614029 4847 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614034 4847 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614039 4847 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614044 4847 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614049 4847 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614054 4847 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614059 4847 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614064 4847 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614069 4847 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614074 4847 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614079 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614084 4847 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614090 4847 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614096 4847 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614109 4847 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614119 4847 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614126 4847 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614133 4847 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614139 4847 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614146 4847 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614152 4847 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614158 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614164 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614173 4847 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614184 4847 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614191 4847 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614199 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614206 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614213 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614220 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614226 4847 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614231 4847 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614236 4847 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614241 4847 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614246 4847 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614251 4847 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614258 4847 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614265 4847 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614278 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614289 4847 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614296 4847 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614302 4847 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614308 4847 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614315 4847 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614321 4847 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614328 4847 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614336 4847 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614345 4847 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614351 4847 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614358 4847 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614364 4847 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614370 4847 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614376 4847 feature_gate.go:330] unrecognized feature gate: Example Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614382 4847 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614389 4847 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614396 4847 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614403 4847 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614412 4847 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614419 4847 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614426 4847 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614432 4847 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614437 4847 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614442 4847 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.614447 4847 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.614456 4847 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.614706 4847 server.go:940] "Client rotation is on, will bootstrap in background" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.617733 4847 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.617846 4847 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.619045 4847 server.go:997] "Starting client certificate rotation" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.619077 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.619564 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-07 00:59:25.801179282 +0000 UTC Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.619663 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.628268 4847 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.630100 4847 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.630270 4847 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.639747 4847 log.go:25] "Validated CRI v1 runtime API" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.655599 4847 log.go:25] "Validated CRI v1 image API" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.657444 4847 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.659979 4847 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-10-14-19-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.660083 4847 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.690196 4847 manager.go:217] Machine: {Timestamp:2025-12-10 14:24:10.686431052 +0000 UTC m=+0.255648762 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7a2b8039-75e9-4d12-a71a-bbfc9770e7fe BootID:ee37509e-2f06-438f-90b6-27523eac2f9d Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:da:b8:ec Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:da:b8:ec Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b5:d4:ad Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:39:9f:de Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:27:42:53 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:33:3f:c2 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fe:76:03:e3:3b:50 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:de:ba:d8:b7:c2 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.690631 4847 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.690902 4847 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.691923 4847 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.692256 4847 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.692311 4847 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.692705 4847 topology_manager.go:138] "Creating topology manager with none policy" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.692758 4847 container_manager_linux.go:303] "Creating device plugin manager" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.693185 4847 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.693256 4847 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.693573 4847 state_mem.go:36] "Initialized new in-memory state store" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.693790 4847 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.696874 4847 kubelet.go:418] "Attempting to sync node with API server" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.696928 4847 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.696963 4847 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.697023 4847 kubelet.go:324] "Adding apiserver pod source" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.697043 4847 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.698623 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.698648 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.698737 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.698740 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.699265 4847 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.699639 4847 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700263 4847 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700814 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700835 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700843 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700850 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700860 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700875 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700882 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700895 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700905 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700913 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700925 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.700932 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.701441 4847 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.701904 4847 server.go:1280] "Started kubelet" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.702063 4847 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.702174 4847 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.702294 4847 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.703158 4847 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 10 14:24:10 crc systemd[1]: Started Kubernetes Kubelet. Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704250 4847 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704288 4847 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704320 4847 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 16:14:51.679620107 +0000 UTC Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704377 4847 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 913h50m40.97524546s for next certificate rotation Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704507 4847 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704515 4847 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.704752 4847 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.705255 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.705318 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.705501 4847 server.go:460] "Adding debug handlers to kubelet server" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.704659 4847 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.707555 4847 factory.go:55] Registering systemd factory Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.707601 4847 factory.go:221] Registration of the systemd container factory successfully Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.708472 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.708786 4847 factory.go:153] Registering CRI-O factory Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.708820 4847 factory.go:221] Registration of the crio container factory successfully Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.708931 4847 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.708967 4847 factory.go:103] Registering Raw factory Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.708999 4847 manager.go:1196] Started watching for new ooms in manager Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.710183 4847 manager.go:319] Starting recovery of all containers Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.711054 4847 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187fe0b5dfda7c7a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 14:24:10.701864058 +0000 UTC m=+0.271081688,LastTimestamp:2025-12-10 14:24:10.701864058 +0000 UTC m=+0.271081688,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720166 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720215 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720227 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720236 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720247 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720257 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720266 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720276 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720286 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720295 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720326 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720337 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720346 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720356 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720366 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720443 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720569 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720579 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720587 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720596 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720607 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720620 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720633 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720644 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720656 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720665 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720677 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720687 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720696 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720728 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720758 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720769 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720780 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720790 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720801 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720814 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720827 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720839 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720853 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720864 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720872 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720881 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720890 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720901 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720913 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720926 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720939 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720951 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720964 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720977 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720985 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.720996 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721014 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721028 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721043 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721055 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721066 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721077 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721097 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721110 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721122 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721133 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721143 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721154 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721165 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721177 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721187 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721197 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721209 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721220 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721231 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721241 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721254 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721265 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721276 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721287 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721300 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721311 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721321 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.721332 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.722913 4847 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.722952 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.722966 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.722979 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.722991 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723003 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723014 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723044 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723056 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723068 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723079 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723090 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723103 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723115 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723127 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723139 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723152 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723175 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723187 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723200 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723213 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723226 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723237 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723249 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723262 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723312 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723327 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723341 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723355 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723369 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723382 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723396 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723410 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723423 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723436 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723448 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723460 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723472 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723485 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723497 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723508 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723518 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723529 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723541 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723551 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723561 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723571 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723582 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723593 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723601 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723610 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723618 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723626 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723635 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723644 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723652 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723661 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723670 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723679 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723688 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723696 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723705 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723737 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723746 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723756 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723764 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723773 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723782 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723790 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723798 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723806 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723815 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723823 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723831 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723840 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723850 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723859 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723867 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723876 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723885 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723893 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723902 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723910 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723918 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723926 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723937 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723946 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723955 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723966 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723975 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723983 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.723992 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724001 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724010 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724019 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724028 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724037 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724046 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724054 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724063 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724099 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724108 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724116 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724124 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724132 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724140 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724148 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724157 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724165 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724173 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724181 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724189 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724198 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724206 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724214 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724223 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724231 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724240 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724248 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724258 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724268 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724279 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724290 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724301 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724311 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724322 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724333 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724342 4847 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724355 4847 reconstruct.go:97] "Volume reconstruction finished" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.724362 4847 reconciler.go:26] "Reconciler: start to sync state" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.733036 4847 manager.go:324] Recovery completed Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.744297 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.745986 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.746136 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.746206 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.747267 4847 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.747287 4847 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.747308 4847 state_mem.go:36] "Initialized new in-memory state store" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.755512 4847 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.757188 4847 policy_none.go:49] "None policy: Start" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.758146 4847 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.758177 4847 state_mem.go:35] "Initializing new in-memory state store" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.758230 4847 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.758272 4847 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.758303 4847 kubelet.go:2335] "Starting kubelet main sync loop" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.758445 4847 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 10 14:24:10 crc kubenswrapper[4847]: W1210 14:24:10.759754 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.759831 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.807307 4847 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.834619 4847 manager.go:334] "Starting Device Plugin manager" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.834689 4847 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.834708 4847 server.go:79] "Starting device plugin registration server" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.835755 4847 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.836043 4847 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.836757 4847 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.837086 4847 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.837106 4847 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.844290 4847 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.858624 4847 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.858801 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.860908 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.860962 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.860975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.861159 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.861497 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.861554 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862394 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862448 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862460 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862529 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862703 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862752 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.862773 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863428 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863567 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863608 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863725 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863788 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.863814 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864592 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864613 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864628 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864644 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864631 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.864839 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865137 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865213 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865589 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865636 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865893 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.865973 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.866251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.866293 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.866308 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.867147 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.867188 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.867200 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.910095 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926467 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926542 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926563 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926584 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926601 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926618 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926633 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926656 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926781 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926832 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926856 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926878 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926902 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926920 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.926938 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.936474 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.937541 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.937640 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.937650 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:10 crc kubenswrapper[4847]: I1210 14:24:10.937671 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:10 crc kubenswrapper[4847]: E1210 14:24:10.938215 4847 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028213 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028261 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028281 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028301 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028319 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028335 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028350 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028384 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028408 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028455 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028467 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028429 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028394 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028420 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028603 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028677 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028700 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028698 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028756 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028760 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028750 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028740 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028804 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028738 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028765 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028869 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028742 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028903 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.028922 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.029019 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.139326 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.140965 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.141033 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.141050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.141083 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:11 crc kubenswrapper[4847]: E1210 14:24:11.141682 4847 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.191082 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.197408 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.218313 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.219322 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6ddbe168497aee7baeeb0e7efd31909087b8b878a5322a34f46eb78a623bebe3 WatchSource:0}: Error finding container 6ddbe168497aee7baeeb0e7efd31909087b8b878a5322a34f46eb78a623bebe3: Status 404 returned error can't find the container with id 6ddbe168497aee7baeeb0e7efd31909087b8b878a5322a34f46eb78a623bebe3 Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.221175 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2ce6023baf366de6960e1723594ee15275b836351c7ef3f733bf4e9b77664512 WatchSource:0}: Error finding container 2ce6023baf366de6960e1723594ee15275b836351c7ef3f733bf4e9b77664512: Status 404 returned error can't find the container with id 2ce6023baf366de6960e1723594ee15275b836351c7ef3f733bf4e9b77664512 Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.231970 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5cdaa4fb0c6e7d20b2f8b84020162d1f4dce912fd2067bdcc37838db81b10687 WatchSource:0}: Error finding container 5cdaa4fb0c6e7d20b2f8b84020162d1f4dce912fd2067bdcc37838db81b10687: Status 404 returned error can't find the container with id 5cdaa4fb0c6e7d20b2f8b84020162d1f4dce912fd2067bdcc37838db81b10687 Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.246409 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.252112 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.263177 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3ea18ff1fa8828cc63a2d57c82b0841c7a2cc87feb803a0855f22ef50585f62d WatchSource:0}: Error finding container 3ea18ff1fa8828cc63a2d57c82b0841c7a2cc87feb803a0855f22ef50585f62d: Status 404 returned error can't find the container with id 3ea18ff1fa8828cc63a2d57c82b0841c7a2cc87feb803a0855f22ef50585f62d Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.269213 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ef8862c1b2a14f06cba4ad24625a156cb18fc83eb03338f2d9c74916e895cd2d WatchSource:0}: Error finding container ef8862c1b2a14f06cba4ad24625a156cb18fc83eb03338f2d9c74916e895cd2d: Status 404 returned error can't find the container with id ef8862c1b2a14f06cba4ad24625a156cb18fc83eb03338f2d9c74916e895cd2d Dec 10 14:24:11 crc kubenswrapper[4847]: E1210 14:24:11.311066 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.541796 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.542998 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.543031 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.543042 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.543068 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:11 crc kubenswrapper[4847]: E1210 14:24:11.543479 4847 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.703416 4847 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.764874 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011" exitCode=0 Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.764952 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.765031 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ef8862c1b2a14f06cba4ad24625a156cb18fc83eb03338f2d9c74916e895cd2d"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.765121 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.767922 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.767967 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.767977 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.768233 4847 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45" exitCode=0 Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.768296 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.768319 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3ea18ff1fa8828cc63a2d57c82b0841c7a2cc87feb803a0855f22ef50585f62d"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.768415 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.769239 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.769273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.769286 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.769670 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770467 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770494 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770504 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770693 4847 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3" exitCode=0 Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770795 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770820 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5cdaa4fb0c6e7d20b2f8b84020162d1f4dce912fd2067bdcc37838db81b10687"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.770886 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.771505 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.771537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.771551 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.773336 4847 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc" exitCode=0 Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.773405 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.773448 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6ddbe168497aee7baeeb0e7efd31909087b8b878a5322a34f46eb78a623bebe3"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.773535 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.774394 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.774434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.774448 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.775821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492"} Dec 10 14:24:11 crc kubenswrapper[4847]: I1210 14:24:11.775850 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2ce6023baf366de6960e1723594ee15275b836351c7ef3f733bf4e9b77664512"} Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.890050 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:11 crc kubenswrapper[4847]: E1210 14:24:11.890125 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:11 crc kubenswrapper[4847]: W1210 14:24:11.963802 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:11 crc kubenswrapper[4847]: E1210 14:24:11.963900 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:12 crc kubenswrapper[4847]: W1210 14:24:12.079084 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:12 crc kubenswrapper[4847]: E1210 14:24:12.079157 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:12 crc kubenswrapper[4847]: E1210 14:24:12.112178 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 10 14:24:12 crc kubenswrapper[4847]: W1210 14:24:12.206074 4847 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 10 14:24:12 crc kubenswrapper[4847]: E1210 14:24:12.206161 4847 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.343738 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.347315 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.347374 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.347385 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.347418 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:12 crc kubenswrapper[4847]: E1210 14:24:12.348156 4847 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.632312 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.780046 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.780098 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.780110 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.780216 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.781056 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.781084 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.781095 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.782836 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.782859 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.782872 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.782938 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.788460 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.788488 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.788499 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791608 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791642 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791655 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791666 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791676 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.791776 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.792543 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.792567 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.792577 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.794098 4847 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276" exitCode=0 Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.794152 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.794239 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.795022 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.795051 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.795060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.795382 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20"} Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.795447 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.796146 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.796165 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:12 crc kubenswrapper[4847]: I1210 14:24:12.796174 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.159198 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.535744 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.538912 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801178 4847 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8" exitCode=0 Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801282 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8"} Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801318 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801355 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801487 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.801496 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802289 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802317 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802291 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802343 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802324 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802409 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802424 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.802440 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.948334 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.949805 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.949836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.949846 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:13 crc kubenswrapper[4847]: I1210 14:24:13.949872 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.813945 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee"} Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc"} Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814020 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814100 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814033 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2"} Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814216 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9"} Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814246 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f"} Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.814170 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815351 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815407 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815353 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815426 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.815458 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.816049 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.816078 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:14 crc kubenswrapper[4847]: I1210 14:24:14.816090 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:15 crc kubenswrapper[4847]: I1210 14:24:15.816989 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:15 crc kubenswrapper[4847]: I1210 14:24:15.818570 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:15 crc kubenswrapper[4847]: I1210 14:24:15.818638 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:15 crc kubenswrapper[4847]: I1210 14:24:15.818662 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:15 crc kubenswrapper[4847]: I1210 14:24:15.997948 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.223936 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.224140 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.226434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.226483 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.226494 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.819999 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.820798 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.820830 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:16 crc kubenswrapper[4847]: I1210 14:24:16.820840 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.389788 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.390592 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.393295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.393336 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.393349 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.394984 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.539663 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.825292 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.826382 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.826461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:18 crc kubenswrapper[4847]: I1210 14:24:18.826500 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:19 crc kubenswrapper[4847]: I1210 14:24:19.827395 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:19 crc kubenswrapper[4847]: I1210 14:24:19.828399 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:19 crc kubenswrapper[4847]: I1210 14:24:19.828469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:19 crc kubenswrapper[4847]: I1210 14:24:19.828486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:20 crc kubenswrapper[4847]: E1210 14:24:20.844646 4847 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.110160 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.110360 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.111571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.111637 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.111652 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.540581 4847 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 14:24:21 crc kubenswrapper[4847]: I1210 14:24:21.540665 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:24:22 crc kubenswrapper[4847]: E1210 14:24:22.633543 4847 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.688983 4847 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.689067 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.703508 4847 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.765202 4847 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.765269 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.771953 4847 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 14:24:22 crc kubenswrapper[4847]: I1210 14:24:22.772020 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.165168 4847 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]log ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]etcd ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/priority-and-fairness-filter ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-apiextensions-informers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-apiextensions-controllers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/crd-informer-synced ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-system-namespaces-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 10 14:24:23 crc kubenswrapper[4847]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/bootstrap-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/start-kube-aggregator-informers ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-registration-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-discovery-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]autoregister-completion ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-openapi-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 10 14:24:23 crc kubenswrapper[4847]: livez check failed Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.166626 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.543328 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.543469 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.544538 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.544569 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:23 crc kubenswrapper[4847]: I1210 14:24:23.544577 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.327596 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.327835 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.328942 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.328981 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.328991 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.363626 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.839700 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.841065 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.841125 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.841156 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:24 crc kubenswrapper[4847]: I1210 14:24:24.859679 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 10 14:24:25 crc kubenswrapper[4847]: I1210 14:24:25.841296 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:25 crc kubenswrapper[4847]: I1210 14:24:25.842156 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:25 crc kubenswrapper[4847]: I1210 14:24:25.842197 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:25 crc kubenswrapper[4847]: I1210 14:24:25.842206 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:26 crc kubenswrapper[4847]: I1210 14:24:26.948525 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 14:24:26 crc kubenswrapper[4847]: I1210 14:24:26.965645 4847 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 10 14:24:27 crc kubenswrapper[4847]: E1210 14:24:27.735747 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.735955 4847 trace.go:236] Trace[850865002]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 14:24:14.992) (total time: 12742ms): Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[850865002]: ---"Objects listed" error: 12742ms (14:24:27.735) Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[850865002]: [12.74295008s] [12.74295008s] END Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.735985 4847 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.736117 4847 trace.go:236] Trace[37013734]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 14:24:14.384) (total time: 13351ms): Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[37013734]: ---"Objects listed" error: 13351ms (14:24:27.736) Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[37013734]: [13.351861576s] [13.351861576s] END Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.736143 4847 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 14:24:27 crc kubenswrapper[4847]: E1210 14:24:27.737902 4847 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.738083 4847 trace.go:236] Trace[217549446]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 14:24:14.366) (total time: 13371ms): Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[217549446]: ---"Objects listed" error: 13371ms (14:24:27.738) Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[217549446]: [13.371456084s] [13.371456084s] END Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.738116 4847 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.738124 4847 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.738641 4847 trace.go:236] Trace[982363352]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 14:24:14.513) (total time: 13224ms): Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[982363352]: ---"Objects listed" error: 13224ms (14:24:27.738) Dec 10 14:24:27 crc kubenswrapper[4847]: Trace[982363352]: [13.224825936s] [13.224825936s] END Dec 10 14:24:27 crc kubenswrapper[4847]: I1210 14:24:27.738668 4847 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.164290 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.169575 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.342964 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.543656 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.547648 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.711142 4847 apiserver.go:52] "Watching apiserver" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.718606 4847 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.719938 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.720359 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.720551 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.720610 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.720670 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.720930 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.721042 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.721103 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.721303 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.721579 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.723653 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.724488 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.724505 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.725606 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.726143 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.726162 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.726221 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.726326 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.726623 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.753242 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.777166 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.787554 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.796739 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.805609 4847 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.812573 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.829959 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.839886 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842825 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842886 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842913 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842936 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842960 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.842982 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843003 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843025 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843045 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843068 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843091 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843114 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843134 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843154 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843175 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843230 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843287 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843329 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843351 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843362 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843373 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843443 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843467 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843490 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843476 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843511 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843526 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843593 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843623 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843623 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843649 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843673 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843697 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843743 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843769 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843938 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.843967 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844241 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.844309 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:24:29.34429178 +0000 UTC m=+18.913509410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844326 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844329 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844369 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844402 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844446 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844471 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844495 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844518 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844562 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844769 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844809 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844808 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844808 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.844994 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.845286 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.845296 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.845406 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848068 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848110 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848152 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848156 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848169 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848217 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848245 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848266 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848335 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848492 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848501 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848522 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848625 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.848648 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849032 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849067 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849090 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849110 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849133 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849154 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849167 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849173 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849221 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849238 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849254 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849269 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849284 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849298 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849313 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849328 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849655 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849689 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849704 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849733 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849749 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849768 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849777 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849783 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849821 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.849995 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850021 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850026 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850044 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850066 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850087 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850109 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850130 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850151 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850172 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850174 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850200 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.850223 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853229 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853319 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853396 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853480 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853565 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853641 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853741 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853823 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854044 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854126 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854199 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854271 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854540 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854614 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854677 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854772 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854867 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854942 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855008 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855095 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855185 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855260 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855332 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855401 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855472 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855542 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855618 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855708 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855843 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855926 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855995 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856074 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856203 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856319 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856407 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856483 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856558 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856637 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856726 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856809 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856886 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856955 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857034 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857115 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857183 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857256 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857326 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857430 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857531 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857611 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857686 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857779 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857861 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857927 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858000 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858106 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858203 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858344 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858421 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858497 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858598 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858682 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858850 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858928 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859089 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859844 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859882 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853078 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853114 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853358 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853431 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853664 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.853918 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860558 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860609 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860685 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860737 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860767 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860796 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860826 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860854 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860874 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860897 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860918 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860942 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860962 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860983 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861007 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861025 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861045 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861076 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861096 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861117 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861140 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861159 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861180 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861200 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861219 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861237 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861256 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861277 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861300 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861318 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861338 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861357 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861374 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861397 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861418 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861434 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861458 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861478 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861497 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861515 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861537 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861557 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861574 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861596 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861629 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861651 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.861671 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854099 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854533 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854674 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.854936 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855053 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855205 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855296 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855330 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855628 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.855590 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856061 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856092 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856324 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856105 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856399 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856402 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856450 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856620 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856770 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856791 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856944 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856975 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.856995 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857001 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857153 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857330 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857381 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857427 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857432 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857650 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858116 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858200 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858207 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.857487 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858273 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858373 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858419 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858433 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858693 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858828 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859034 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858573 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859061 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859260 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859503 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859523 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859528 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859538 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859632 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.859911 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860150 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860224 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860244 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860251 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860353 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860406 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.860497 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863003 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.858063 4847 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863158 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863184 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863260 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863262 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863420 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863460 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863581 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864514 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863664 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863815 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.863969 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864007 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864098 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864239 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864338 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864350 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864346 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864756 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864787 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864808 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864826 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864841 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864856 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864871 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864886 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864902 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864917 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864934 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864949 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.864989 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865010 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865026 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865043 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865058 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865078 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865095 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865111 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865127 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865146 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865162 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865178 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865194 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865216 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865468 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865479 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865488 4847 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865497 4847 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865507 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865517 4847 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865525 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865534 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865542 4847 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865550 4847 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865558 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865567 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865575 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865583 4847 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865592 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865601 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865610 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865619 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865627 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865636 4847 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865644 4847 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865654 4847 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865663 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865672 4847 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865681 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865689 4847 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865698 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865707 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865772 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865781 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865789 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865799 4847 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865808 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865817 4847 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865825 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865834 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865844 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865853 4847 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865861 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865870 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865926 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865937 4847 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865947 4847 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865946 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865955 4847 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.865984 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866005 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866153 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866173 4847 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866176 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866185 4847 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866264 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866283 4847 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866306 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866320 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866334 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866347 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866362 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866381 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866390 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866443 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866526 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866631 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866648 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867033 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.866395 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867145 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867159 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867171 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867316 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.858501 4847 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867345 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867357 4847 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867440 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867454 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867465 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867532 4847 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867641 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867669 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867682 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867770 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867908 4847 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867921 4847 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867930 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867938 4847 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.867990 4847 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868003 4847 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868012 4847 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868022 4847 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868031 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868040 4847 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868048 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868056 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868064 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868072 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868081 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868089 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868097 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868136 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868144 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868153 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868161 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868169 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868177 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868185 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868193 4847 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868202 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868210 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868219 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868227 4847 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868235 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868243 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868251 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868260 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868269 4847 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868277 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868285 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868293 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868301 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868310 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868318 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868326 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868336 4847 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868348 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868358 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868370 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868381 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868392 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868402 4847 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868422 4847 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868431 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868439 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868448 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.868457 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.858439 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.869514 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.869896 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870000 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870087 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870227 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870264 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870270 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870493 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870516 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870589 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870602 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.870840 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871077 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871118 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871212 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871460 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871680 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871779 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871776 4847 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871960 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.871486 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.872403 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.872639 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.872972 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.873201 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.873430 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.873702 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.874054 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.874371 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.874660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.874860 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.875018 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.875104 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:29.37508636 +0000 UTC m=+18.944304080 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.875201 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.875494 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.875703 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.875788 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.875994 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.876047 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.876265 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.876288 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.876565 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.876627 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.875668 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.877032 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.877060 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.876774 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:29.375523242 +0000 UTC m=+18.944740962 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.877321 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.878074 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.878257 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.878837 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.880132 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.878249 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.881132 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.881935 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.881990 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.884534 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.888391 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.888629 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.888774 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.888875 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.888900 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.889145 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.889675 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.890222 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.890337 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.890404 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.893139 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.893317 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.893970 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894000 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894017 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894077 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:29.394059529 +0000 UTC m=+18.963277229 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.893975 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894257 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894268 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:28 crc kubenswrapper[4847]: E1210 14:24:28.894300 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:29.394290535 +0000 UTC m=+18.963508255 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.894829 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.899921 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.900899 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.902446 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.909456 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.912972 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.918238 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968795 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968879 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968920 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968935 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968948 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968962 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968973 4847 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968983 4847 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.968994 4847 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969007 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969018 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969029 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969040 4847 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969051 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969062 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969073 4847 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969083 4847 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969093 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969103 4847 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969099 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969114 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969156 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969181 4847 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969243 4847 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969274 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969299 4847 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969323 4847 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969346 4847 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969370 4847 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969396 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969419 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969442 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969465 4847 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969489 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969512 4847 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969534 4847 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969557 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969580 4847 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969604 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969627 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969651 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969673 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969698 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969768 4847 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969792 4847 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969816 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969839 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969863 4847 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969889 4847 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969911 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969934 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969960 4847 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.969984 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970007 4847 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970022 4847 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970038 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970054 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970073 4847 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970091 4847 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970109 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970126 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970143 4847 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970159 4847 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970175 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970192 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970208 4847 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970223 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970238 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970253 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970269 4847 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970286 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970303 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970319 4847 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970337 4847 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970354 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970369 4847 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970386 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970401 4847 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:28 crc kubenswrapper[4847]: I1210 14:24:28.970416 4847 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.036986 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.050244 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.056673 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 14:24:29 crc kubenswrapper[4847]: W1210 14:24:29.061380 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ba56936b6f400811ec69ca292c6fa8aa381f635a5100ad28d9cb5738b89b8de1 WatchSource:0}: Error finding container ba56936b6f400811ec69ca292c6fa8aa381f635a5100ad28d9cb5738b89b8de1: Status 404 returned error can't find the container with id ba56936b6f400811ec69ca292c6fa8aa381f635a5100ad28d9cb5738b89b8de1 Dec 10 14:24:29 crc kubenswrapper[4847]: W1210 14:24:29.072508 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a2426e923efd7801bbc3ebcf24d84346db298f9721a295426aab8c4283cfffee WatchSource:0}: Error finding container a2426e923efd7801bbc3ebcf24d84346db298f9721a295426aab8c4283cfffee: Status 404 returned error can't find the container with id a2426e923efd7801bbc3ebcf24d84346db298f9721a295426aab8c4283cfffee Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.373023 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.373198 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:24:30.373179651 +0000 UTC m=+19.942397281 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.473538 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.473589 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.473612 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.473637 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473678 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473753 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473769 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473794 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473810 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473773 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473856 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473759 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:30.473741597 +0000 UTC m=+20.042959227 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473895 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:30.473883701 +0000 UTC m=+20.043101341 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473910 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:30.473902852 +0000 UTC m=+20.043120492 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.473940 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: E1210 14:24:29.474066 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:30.474048486 +0000 UTC m=+20.043266116 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.852517 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.852565 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2f03f65343d74acdfec0dddc2206490b2653b027c056530a31f67e6fc7d954a3"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.853656 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a2426e923efd7801bbc3ebcf24d84346db298f9721a295426aab8c4283cfffee"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.855934 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.855976 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.855990 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ba56936b6f400811ec69ca292c6fa8aa381f635a5100ad28d9cb5738b89b8de1"} Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.868581 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.880310 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.894294 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.909935 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.938986 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.959510 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.979251 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:29 crc kubenswrapper[4847]: I1210 14:24:29.994000 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.007638 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.022801 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.042187 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.065186 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.094762 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.111324 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.123844 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.138453 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.382216 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.382449 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:24:32.382421269 +0000 UTC m=+21.951638899 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.482910 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.482973 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.483000 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483018 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483089 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483110 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:32.483090168 +0000 UTC m=+22.052307808 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483114 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483130 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.483025 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483138 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483191 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483207 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483164 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:32.48315308 +0000 UTC m=+22.052370710 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483246 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483270 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:32.483254693 +0000 UTC m=+22.052472323 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.483381 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:32.483359455 +0000 UTC m=+22.052577085 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.759377 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.759523 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.759571 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.759691 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.759808 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.759986 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.763869 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.764452 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.765754 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.766327 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.767279 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.767795 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.768322 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.769360 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.769997 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.770916 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.771447 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.772468 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.772987 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.773479 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.774045 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.774383 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.775072 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.776051 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.776430 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.776982 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.777992 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.778461 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.779405 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.779957 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.780993 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.781415 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.782010 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.783137 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.783596 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.784533 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.785016 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.785846 4847 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.785952 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.787771 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.788740 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.789221 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.790477 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.790864 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.791582 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.792448 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.793168 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.794240 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.794756 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.795861 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.796592 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.797677 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.798332 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.799381 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.799875 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.801036 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.801525 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.802783 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.803916 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.805959 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.806887 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.807552 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.811575 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.827443 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.848296 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.863482 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.882407 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.899684 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.938074 4847 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.941216 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.941261 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.941273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.941386 4847 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.952319 4847 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.952825 4847 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.954133 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.954181 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.954194 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.954213 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.954226 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:30Z","lastTransitionTime":"2025-12-10T14:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.978740 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.982779 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.982831 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.982844 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.982865 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:30 crc kubenswrapper[4847]: I1210 14:24:30.982879 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:30Z","lastTransitionTime":"2025-12-10T14:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:30 crc kubenswrapper[4847]: E1210 14:24:30.998743 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.002435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.002474 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.002483 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.002498 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.002508 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: E1210 14:24:31.016753 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.021281 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.021341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.021353 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.021373 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.021392 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: E1210 14:24:31.043313 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.047305 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.047367 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.047378 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.047394 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.047406 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: E1210 14:24:31.066259 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: E1210 14:24:31.066466 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.069226 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.069273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.069284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.069307 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.069318 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.171678 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.171733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.171743 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.171761 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.171772 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.256533 4847 csr.go:261] certificate signing request csr-htrrb is approved, waiting to be issued Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.270367 4847 csr.go:257] certificate signing request csr-htrrb is issued Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.274517 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.274572 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.274583 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.274607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.274620 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.377060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.377123 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.377136 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.377165 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.377185 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.479966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.480030 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.480041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.480063 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.480405 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.583241 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.583306 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.583316 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.583337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.583367 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.685463 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.685503 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.685512 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.685526 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.685536 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.787941 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.787977 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.787987 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.788004 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.788016 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.863085 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.887731 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.890332 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.890375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.890387 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.890403 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.890420 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.918395 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.942594 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.976511 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.992437 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.993386 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.993456 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.993476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.993508 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:31 crc kubenswrapper[4847]: I1210 14:24:31.993526 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:31Z","lastTransitionTime":"2025-12-10T14:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.014882 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.038599 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.054779 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.095549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.095610 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.095619 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.095649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.095674 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.112739 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gvdwq"] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.113145 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-v58hl"] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.113376 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.113769 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.120733 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b5l5q"] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.120873 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.121023 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.121023 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.121276 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.121784 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.121983 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.122017 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.122126 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-v9dh5"] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.122601 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.122942 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.122974 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.123083 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.123141 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-67b5l"] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.123399 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.124626 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.124896 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.126470 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.126665 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.126886 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.127072 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.127301 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.127521 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.127564 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.127576 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.128143 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.128800 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.129500 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.135317 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.152219 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.166467 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.181158 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.195251 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.197862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.197909 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.197921 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.197953 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.197966 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.200747 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-mcd-auth-proxy-config\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.200777 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.200794 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.200825 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.200925 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfmr2\" (UniqueName: \"kubernetes.io/projected/8ab74b21-e521-4884-9b54-207201f6b78f-kube-api-access-dfmr2\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201016 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-proxy-tls\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201059 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201084 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-kubelet\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201117 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-system-cni-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201150 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cnibin\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201174 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-binary-copy\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201200 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-system-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201227 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-os-release\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201259 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201291 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-k8s-cni-cncf-io\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201314 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-bin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201338 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201360 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201377 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201399 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhhpv\" (UniqueName: \"kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201427 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-multus\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201462 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201557 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-cnibin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201630 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201655 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4bm\" (UniqueName: \"kubernetes.io/projected/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-kube-api-access-gj4bm\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201692 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-conf-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201764 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-daemon-config\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201824 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqxps\" (UniqueName: \"kubernetes.io/projected/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-kube-api-access-pqxps\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201851 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201895 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201916 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-os-release\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201938 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.201958 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-hostroot\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202000 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-multus-certs\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202021 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8ab74b21-e521-4884-9b54-207201f6b78f-hosts-file\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202041 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwjkr\" (UniqueName: \"kubernetes.io/projected/71681159-7da8-4bc9-837c-d0e3b7397e2e-kube-api-access-jwjkr\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202100 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-rootfs\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202122 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202141 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202163 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202181 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-etc-kubernetes\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202223 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202247 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-socket-dir-parent\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202302 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202324 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202372 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202404 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202444 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202468 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-cni-binary-copy\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.202496 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-netns\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.208333 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.219070 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.231725 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.242740 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.256513 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.267396 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.272497 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-10 14:19:31 +0000 UTC, rotation deadline is 2026-09-24 12:22:04.685251097 +0000 UTC Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.272564 4847 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6909h57m32.412691165s for next certificate rotation Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.276328 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.288592 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.300384 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.300436 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.300451 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.300469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.300481 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.302957 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303092 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqxps\" (UniqueName: \"kubernetes.io/projected/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-kube-api-access-pqxps\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303136 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303159 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303180 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-os-release\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303201 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303222 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-hostroot\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303246 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-multus-certs\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303268 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8ab74b21-e521-4884-9b54-207201f6b78f-hosts-file\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303289 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwjkr\" (UniqueName: \"kubernetes.io/projected/71681159-7da8-4bc9-837c-d0e3b7397e2e-kube-api-access-jwjkr\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303300 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303341 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-os-release\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303312 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-rootfs\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303287 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303354 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-multus-certs\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303382 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8ab74b21-e521-4884-9b54-207201f6b78f-hosts-file\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303393 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303443 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303450 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-hostroot\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303484 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303463 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303509 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-etc-kubernetes\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303529 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-etc-kubernetes\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303343 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-rootfs\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303441 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303558 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303576 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303615 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-socket-dir-parent\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303666 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-socket-dir-parent\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303672 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303765 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303819 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303868 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303918 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.303963 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-cni-binary-copy\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304005 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-netns\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304075 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-mcd-auth-proxy-config\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304095 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304109 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304123 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304165 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304084 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304199 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304236 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304252 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-netns\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304268 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304298 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304379 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304388 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304421 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304438 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfmr2\" (UniqueName: \"kubernetes.io/projected/8ab74b21-e521-4884-9b54-207201f6b78f-kube-api-access-dfmr2\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304461 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304486 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-proxy-tls\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304575 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-kubelet\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304620 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-system-cni-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304675 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304678 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cnibin\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304760 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-binary-copy\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304816 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-system-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304858 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-os-release\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304877 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-mcd-auth-proxy-config\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304909 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304924 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cnibin\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304954 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-kubelet\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304961 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-system-cni-dir\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304958 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-k8s-cni-cncf-io\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305007 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-run-k8s-cni-cncf-io\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305010 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305029 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-bin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.304891 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-cni-binary-copy\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305064 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-os-release\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305142 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305180 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-bin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305184 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-system-cni-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305204 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305230 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305253 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhhpv\" (UniqueName: \"kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305277 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305359 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-multus\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305376 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-host-var-lib-cni-multus\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305585 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-cni-binary-copy\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305626 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305726 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-cnibin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305763 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305779 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4bm\" (UniqueName: \"kubernetes.io/projected/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-kube-api-access-gj4bm\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305798 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-conf-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305817 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-daemon-config\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.306019 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.305693 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.306183 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-cnibin\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.306224 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-conf-dir\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.306260 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/71681159-7da8-4bc9-837c-d0e3b7397e2e-multus-daemon-config\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.306847 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.316337 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-proxy-tls\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.316431 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.320747 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfmr2\" (UniqueName: \"kubernetes.io/projected/8ab74b21-e521-4884-9b54-207201f6b78f-kube-api-access-dfmr2\") pod \"node-resolver-v9dh5\" (UID: \"8ab74b21-e521-4884-9b54-207201f6b78f\") " pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.324994 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwjkr\" (UniqueName: \"kubernetes.io/projected/71681159-7da8-4bc9-837c-d0e3b7397e2e-kube-api-access-jwjkr\") pod \"multus-67b5l\" (UID: \"71681159-7da8-4bc9-837c-d0e3b7397e2e\") " pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.327179 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqxps\" (UniqueName: \"kubernetes.io/projected/8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324-kube-api-access-pqxps\") pod \"machine-config-daemon-gvdwq\" (UID: \"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\") " pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.329482 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4bm\" (UniqueName: \"kubernetes.io/projected/550fdaaf-ffec-4047-b02d-3002d3fc3e2b-kube-api-access-gj4bm\") pod \"multus-additional-cni-plugins-v58hl\" (UID: \"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\") " pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.330656 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.330887 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhhpv\" (UniqueName: \"kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv\") pod \"ovnkube-node-b5l5q\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.349638 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.364965 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.387294 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.401972 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.402851 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.402885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.402894 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.402909 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.402921 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.406585 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.406691 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:24:36.406668754 +0000 UTC m=+25.975886394 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.413405 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.427812 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.428188 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.436659 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v58hl" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.445355 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.448559 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: W1210 14:24:32.450497 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod550fdaaf_ffec_4047_b02d_3002d3fc3e2b.slice/crio-feeb8062c6702ed746a60bcfa1fdccc21d8c455651ca817e18b1d81b4dfe3919 WatchSource:0}: Error finding container feeb8062c6702ed746a60bcfa1fdccc21d8c455651ca817e18b1d81b4dfe3919: Status 404 returned error can't find the container with id feeb8062c6702ed746a60bcfa1fdccc21d8c455651ca817e18b1d81b4dfe3919 Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.452921 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v9dh5" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.461845 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-67b5l" Dec 10 14:24:32 crc kubenswrapper[4847]: W1210 14:24:32.465785 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b405b3d_ae75_409a_9e83_b098f333a5c0.slice/crio-a1efc85f1dd13bc6a32c659829e3afc2b08af6502658d64d2236aa23de9dd8fd WatchSource:0}: Error finding container a1efc85f1dd13bc6a32c659829e3afc2b08af6502658d64d2236aa23de9dd8fd: Status 404 returned error can't find the container with id a1efc85f1dd13bc6a32c659829e3afc2b08af6502658d64d2236aa23de9dd8fd Dec 10 14:24:32 crc kubenswrapper[4847]: W1210 14:24:32.467327 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab74b21_e521_4884_9b54_207201f6b78f.slice/crio-3d879d3bff1f1aec0be9cc4a84e575b3382bb5a85520cc2452814c7f603f50fd WatchSource:0}: Error finding container 3d879d3bff1f1aec0be9cc4a84e575b3382bb5a85520cc2452814c7f603f50fd: Status 404 returned error can't find the container with id 3d879d3bff1f1aec0be9cc4a84e575b3382bb5a85520cc2452814c7f603f50fd Dec 10 14:24:32 crc kubenswrapper[4847]: W1210 14:24:32.489275 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71681159_7da8_4bc9_837c_d0e3b7397e2e.slice/crio-8d1688490434e58c49516834e72d51254632ba24295e0942a1d83497c5c029c5 WatchSource:0}: Error finding container 8d1688490434e58c49516834e72d51254632ba24295e0942a1d83497c5c029c5: Status 404 returned error can't find the container with id 8d1688490434e58c49516834e72d51254632ba24295e0942a1d83497c5c029c5 Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.508416 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.508528 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.508540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.508559 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.508577 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.509570 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.510139 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510143 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.510288 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510354 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:36.510328824 +0000 UTC m=+26.079546454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.510383 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510486 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510585 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:36.510561541 +0000 UTC m=+26.079779171 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510497 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510638 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510659 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510727 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:36.510699404 +0000 UTC m=+26.079917034 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510276 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510805 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510819 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.510862 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:36.510852688 +0000 UTC m=+26.080070318 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.614857 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.614884 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.614893 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.614918 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.614927 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.717335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.718007 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.718026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.718051 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.718074 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.758668 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.758737 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.758829 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.758838 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.758942 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:32 crc kubenswrapper[4847]: E1210 14:24:32.759058 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.820535 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.820585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.820596 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.820613 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.820624 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.868164 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.868223 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.868238 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"98db2c8258dc428e3bac15bdccb488cc7b4e899e5747eded4086c76befe6e009"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.869553 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerStarted","Data":"b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.869588 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerStarted","Data":"8d1688490434e58c49516834e72d51254632ba24295e0942a1d83497c5c029c5"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.870990 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v9dh5" event={"ID":"8ab74b21-e521-4884-9b54-207201f6b78f","Type":"ContainerStarted","Data":"7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.871034 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v9dh5" event={"ID":"8ab74b21-e521-4884-9b54-207201f6b78f","Type":"ContainerStarted","Data":"3d879d3bff1f1aec0be9cc4a84e575b3382bb5a85520cc2452814c7f603f50fd"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.872367 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" exitCode=0 Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.872428 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.872457 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"a1efc85f1dd13bc6a32c659829e3afc2b08af6502658d64d2236aa23de9dd8fd"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.874177 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991" exitCode=0 Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.874272 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.874322 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerStarted","Data":"feeb8062c6702ed746a60bcfa1fdccc21d8c455651ca817e18b1d81b4dfe3919"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.896242 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.923885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.923934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.923946 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.923970 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.923985 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:32Z","lastTransitionTime":"2025-12-10T14:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.931095 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:32 crc kubenswrapper[4847]: I1210 14:24:32.952592 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.003989 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.029773 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.029818 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.029830 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.029848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.029860 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.034283 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.057842 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.079908 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.092579 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.108068 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.125672 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.131988 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.132017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.132026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.132040 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.132049 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.139953 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.152557 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.163390 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.180122 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.198350 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.209122 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.219865 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.231326 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.233810 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.233845 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.233852 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.233865 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.233874 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.241257 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.253028 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.265669 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.276032 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.286362 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.299486 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.311280 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.333642 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.336271 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.336322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.336337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.336359 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.336378 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.439089 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.439123 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.439132 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.439145 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.439155 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.540920 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.540947 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.540955 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.540966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.540976 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.643501 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.643541 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.643552 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.643571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.643583 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.746120 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.746168 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.746180 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.746195 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.746205 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.849114 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.849502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.849515 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.849530 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.849540 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.879425 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.881550 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerStarted","Data":"1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.897605 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.913472 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.929523 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.944337 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.952578 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.952624 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.952634 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.952651 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.952662 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:33Z","lastTransitionTime":"2025-12-10T14:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.960974 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.973744 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:33 crc kubenswrapper[4847]: I1210 14:24:33.989665 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:33Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.008564 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.023729 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.037630 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.051003 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.062089 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.062121 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.062134 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.062150 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.062162 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.074847 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.091202 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.165274 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.165314 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.165323 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.165340 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.165350 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.267018 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.267350 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.267364 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.267385 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.267399 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.324177 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5x4hl"] Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.324579 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.326467 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.327131 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.327198 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.327407 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.340357 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.358484 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.369975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.370015 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.370026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.370044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.370055 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.373654 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.388107 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.402741 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.426327 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.430147 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/252e084b-a363-4110-95c6-57bf80cb730a-serviceca\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.430240 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g7jl\" (UniqueName: \"kubernetes.io/projected/252e084b-a363-4110-95c6-57bf80cb730a-kube-api-access-8g7jl\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.430386 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/252e084b-a363-4110-95c6-57bf80cb730a-host\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.438426 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.451872 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.463116 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.472219 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.472246 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.472256 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.472269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.472278 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.480167 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.492771 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.505475 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.516061 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.524880 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.531200 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/252e084b-a363-4110-95c6-57bf80cb730a-host\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.531340 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/252e084b-a363-4110-95c6-57bf80cb730a-serviceca\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.531369 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g7jl\" (UniqueName: \"kubernetes.io/projected/252e084b-a363-4110-95c6-57bf80cb730a-kube-api-access-8g7jl\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.531371 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/252e084b-a363-4110-95c6-57bf80cb730a-host\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.532354 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/252e084b-a363-4110-95c6-57bf80cb730a-serviceca\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.548941 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g7jl\" (UniqueName: \"kubernetes.io/projected/252e084b-a363-4110-95c6-57bf80cb730a-kube-api-access-8g7jl\") pod \"node-ca-5x4hl\" (UID: \"252e084b-a363-4110-95c6-57bf80cb730a\") " pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.575297 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.575354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.575366 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.575382 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.575741 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.641399 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5x4hl" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.678603 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.678906 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.678992 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.679101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.679123 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.759109 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:34 crc kubenswrapper[4847]: E1210 14:24:34.759222 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.759824 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.759124 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:34 crc kubenswrapper[4847]: E1210 14:24:34.759953 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:34 crc kubenswrapper[4847]: E1210 14:24:34.760030 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.782025 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.782067 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.782078 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.782095 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.782108 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.884134 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.884186 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.884201 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.884218 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.884230 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.886742 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5x4hl" event={"ID":"252e084b-a363-4110-95c6-57bf80cb730a","Type":"ContainerStarted","Data":"8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.886780 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5x4hl" event={"ID":"252e084b-a363-4110-95c6-57bf80cb730a","Type":"ContainerStarted","Data":"7975cedc3777ee0afd4f44178fbd659fcad233107ecbb62f8b64878718d5243a"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.891616 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.891661 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.891675 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.891688 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.891701 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.894778 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7" exitCode=0 Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.894821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7"} Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.899807 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.912051 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.927910 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.942477 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.955161 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.967461 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.980614 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.986654 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.986686 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.986695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.986724 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:34 crc kubenswrapper[4847]: I1210 14:24:34.986733 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:34Z","lastTransitionTime":"2025-12-10T14:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.000146 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:34Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.012923 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.025644 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.039808 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.054017 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.065599 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.077946 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.088814 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.088858 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.088867 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.088882 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.088892 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.090140 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.104058 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.116437 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.128141 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.138498 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.164510 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.186553 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.191223 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.191251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.191259 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.191272 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.191280 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.208914 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.226817 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.238093 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.248152 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.260626 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.270902 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.286525 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.294185 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.294232 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.294243 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.294257 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.294267 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.396824 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.396888 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.396904 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.396923 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.396938 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.499287 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.499341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.499360 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.499382 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.499399 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.602805 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.602893 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.602910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.602934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.602951 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.705218 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.705267 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.705281 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.705301 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.705317 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.808044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.808102 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.808119 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.808136 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.808147 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.900304 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a" exitCode=0 Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.900348 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.910499 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.910536 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.910544 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.910560 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.910570 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:35Z","lastTransitionTime":"2025-12-10T14:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.927431 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.945947 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.961888 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.977797 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:35 crc kubenswrapper[4847]: I1210 14:24:35.992099 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:35Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.010283 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.013307 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.013359 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.013372 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.013393 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.013405 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.025244 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.038756 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.056155 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.070245 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.082234 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.095480 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.108374 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.115513 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.115553 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.115566 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.115581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.115591 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.122779 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.217952 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.217999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.218013 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.218031 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.218043 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.320482 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.320527 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.320537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.320553 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.320564 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.422994 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.423040 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.423053 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.423072 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.423087 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.451008 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.451292 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:24:44.451261996 +0000 UTC m=+34.020479636 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.525405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.525465 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.525479 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.525496 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.525508 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.551971 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.552016 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.552048 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.552079 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552213 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552229 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552224 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552330 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552342 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552381 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552390 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:44.552372816 +0000 UTC m=+34.121590446 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552395 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552450 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:44.552418798 +0000 UTC m=+34.121636428 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552487 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:44.552476579 +0000 UTC m=+34.121694199 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552241 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.552541 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:44.552535021 +0000 UTC m=+34.121752651 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.628357 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.628397 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.628409 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.628423 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.628432 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.730335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.730379 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.730391 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.730406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.730415 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.759262 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.759367 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.759403 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.759436 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.759479 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:36 crc kubenswrapper[4847]: E1210 14:24:36.759571 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.832665 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.832733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.832745 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.832763 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.832776 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.908185 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.911411 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07" exitCode=0 Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.911464 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.924744 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.934647 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.934886 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.934896 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.934910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.934919 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:36Z","lastTransitionTime":"2025-12-10T14:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.942770 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.954040 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.966092 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.980764 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:36 crc kubenswrapper[4847]: I1210 14:24:36.994163 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:36Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.007226 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.023050 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.036354 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.037434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.037471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.037479 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.037493 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.037504 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.046330 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.059201 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.072362 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.086938 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.101442 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.139744 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.139786 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.139797 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.139814 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.139826 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.242080 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.242121 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.242129 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.242143 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.242152 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.345341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.345423 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.345438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.345463 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.345477 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.448034 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.448080 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.448092 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.448109 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.448123 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.551602 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.551640 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.551648 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.551663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.551672 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.653826 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.653867 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.653876 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.653891 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.653901 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.756753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.756798 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.756810 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.756831 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.756843 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.859910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.859945 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.859954 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.859968 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.859977 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.919287 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d" exitCode=0 Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.919332 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.931917 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.947675 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962346 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962827 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962882 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962897 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.962909 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:37Z","lastTransitionTime":"2025-12-10T14:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.977967 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:37 crc kubenswrapper[4847]: I1210 14:24:37.992641 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:37Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.017861 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.029434 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.041766 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.053361 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.065113 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.065190 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.065206 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.065223 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.065235 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.069323 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.081673 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.092744 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.103217 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.113324 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.168997 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.169052 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.169063 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.169085 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.169095 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.271899 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.271961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.271979 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.272003 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.272019 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.374022 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.374061 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.374073 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.374087 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.374096 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.476911 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.476956 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.476967 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.476991 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.477002 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.580070 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.580382 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.580397 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.580410 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.580420 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.683200 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.683237 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.683247 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.683260 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.683270 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.759123 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.759167 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.759253 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:38 crc kubenswrapper[4847]: E1210 14:24:38.759388 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:38 crc kubenswrapper[4847]: E1210 14:24:38.759484 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:38 crc kubenswrapper[4847]: E1210 14:24:38.759566 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.785866 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.785901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.785912 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.785926 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.785936 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.888598 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.888635 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.888647 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.888663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.888673 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.928471 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.928740 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.928763 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.931622 4847 generic.go:334] "Generic (PLEG): container finished" podID="550fdaaf-ffec-4047-b02d-3002d3fc3e2b" containerID="08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47" exitCode=0 Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.931669 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerDied","Data":"08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47"} Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.940813 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.952353 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.956732 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.970499 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.989383 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:38Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.991854 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.991919 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.991928 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.991944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:38 crc kubenswrapper[4847]: I1210 14:24:38.991977 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:38Z","lastTransitionTime":"2025-12-10T14:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.008922 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.026590 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.043292 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.059005 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.077400 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.090445 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.095397 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.095435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.095443 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.095459 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.095469 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.102427 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.117944 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.130862 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.141025 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.152594 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.164997 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.183586 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.195594 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.197406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.197447 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.197455 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.197469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.197478 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.205966 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.217176 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.228064 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.236834 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.248892 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.259770 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.270882 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.282362 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.292662 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.299103 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.299141 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.299151 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.299164 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.299174 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.308294 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.401557 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.401627 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.401649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.401677 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.401700 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.503662 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.503698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.503724 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.503740 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.503751 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.606390 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.606430 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.606439 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.606452 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.606466 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.708249 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.708312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.708323 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.708338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.708371 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.810625 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.810666 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.810676 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.810693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.810705 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.913242 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.913285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.913298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.913313 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.913325 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:39Z","lastTransitionTime":"2025-12-10T14:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.938558 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" event={"ID":"550fdaaf-ffec-4047-b02d-3002d3fc3e2b","Type":"ContainerStarted","Data":"0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4"} Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.939105 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.952888 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.961826 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.964364 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.977863 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.989273 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:39 crc kubenswrapper[4847]: I1210 14:24:39.997421 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:39Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.006367 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.015479 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.015516 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.015527 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.015577 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.015589 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.017919 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.026035 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.038374 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.053459 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.069087 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.087704 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.101495 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.118214 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.118242 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.118255 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.118271 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.118282 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.121608 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.143788 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.189917 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.201343 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.219128 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.220362 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.220403 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.220412 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.220430 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.220440 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.229777 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.240531 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.252405 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.263626 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.281390 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.291429 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.304613 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.315360 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.322937 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.322990 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.323002 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.323021 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.323034 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.331687 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.346512 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.425649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.425698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.425735 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.425769 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.425783 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.528305 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.528569 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.528580 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.528595 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.528606 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.619242 4847 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.630407 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.630652 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.630767 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.630853 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.630935 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.733415 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.733453 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.733463 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.733478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.733487 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.758836 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.758933 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:40 crc kubenswrapper[4847]: E1210 14:24:40.758966 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.758849 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:40 crc kubenswrapper[4847]: E1210 14:24:40.759084 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:40 crc kubenswrapper[4847]: E1210 14:24:40.759151 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.777808 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.792823 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.804216 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.817872 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.831568 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.835169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.835196 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.835207 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.835221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.835232 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.844850 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.856317 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.866479 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.875901 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.891124 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.910248 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.923177 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.934919 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.937055 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.937091 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.937101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.937119 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.937130 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:40Z","lastTransitionTime":"2025-12-10T14:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.943277 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/0.log" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.945439 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a" exitCode=1 Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.945477 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a"} Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.945994 4847 scope.go:117] "RemoveContainer" containerID="16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.949936 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.961392 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.977323 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:40Z\\\",\\\"message\\\":\\\":24:40.651704 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 14:24:40.651746 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 14:24:40.651782 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 14:24:40.651831 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 14:24:40.651866 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 14:24:40.651907 6103 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 14:24:40.652389 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 14:24:40.652411 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 14:24:40.652429 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 14:24:40.652762 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 14:24:40.652887 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 14:24:40.652934 6103 factory.go:656] Stopping watch factory\\\\nI1210 14:24:40.652965 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1210 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.987286 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:40 crc kubenswrapper[4847]: I1210 14:24:40.997165 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.009984 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.020646 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.029960 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.038676 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.038701 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.038724 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.038737 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.038746 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.041507 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.054774 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.067700 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.088673 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.102748 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.117390 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.130462 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.141221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.141273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.141287 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.141306 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.141317 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.244173 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.244210 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.244220 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.244235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.244248 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.346882 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.346928 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.346938 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.346953 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.346965 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.354136 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.354191 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.354205 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.354224 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.354234 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.366410 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.369632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.369662 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.369674 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.369689 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.369699 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.381756 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.384872 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.384919 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.384934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.384950 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.384963 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.400903 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.404582 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.404620 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.404632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.404648 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.404660 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.417073 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.420333 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.420370 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.420381 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.420395 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.420405 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.433382 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.433524 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.448900 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.448931 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.448942 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.448959 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.448972 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.551245 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.551275 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.551283 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.551298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.551308 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.652883 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.652923 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.652933 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.652948 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.652959 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.754695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.754748 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.754756 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.754768 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.754777 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.857614 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.857672 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.857691 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.857740 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.857766 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.950264 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/1.log" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.950984 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/0.log" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.954158 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565" exitCode=1 Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.954203 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.954259 4847 scope.go:117] "RemoveContainer" containerID="16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.955035 4847 scope.go:117] "RemoveContainer" containerID="004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565" Dec 10 14:24:41 crc kubenswrapper[4847]: E1210 14:24:41.955264 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.960646 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.960689 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.960699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.960736 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.960937 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:41Z","lastTransitionTime":"2025-12-10T14:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.976791 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:41 crc kubenswrapper[4847]: I1210 14:24:41.993166 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.006111 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.019604 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.033502 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.053543 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16907fe38d6c641ee99280b719f6573afde3d7b3442a815cfe423d1d7d9b543a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:40Z\\\",\\\"message\\\":\\\":24:40.651704 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 14:24:40.651746 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 14:24:40.651782 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 14:24:40.651831 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 14:24:40.651866 6103 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 14:24:40.651907 6103 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 14:24:40.652389 6103 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 14:24:40.652411 6103 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 14:24:40.652429 6103 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 14:24:40.652762 6103 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 14:24:40.652887 6103 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 14:24:40.652934 6103 factory.go:656] Stopping watch factory\\\\nI1210 14:24:40.652965 6103 ovnkube.go:599] Stopped ovnkube\\\\nI1210 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.063917 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.063960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.063971 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.063987 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.063997 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.067270 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.077377 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.091860 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.104914 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.115013 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.126403 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.136898 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.145687 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.166438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.166476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.166486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.166502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.166512 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.268529 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.268608 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.268623 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.268641 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.268660 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.370704 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.370746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.370754 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.370770 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.370779 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.486684 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.486749 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.486763 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.486790 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.486803 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.589418 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.589458 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.589470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.589486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.589499 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.691777 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.692030 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.692119 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.692223 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.692300 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.759266 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.759333 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:42 crc kubenswrapper[4847]: E1210 14:24:42.759391 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:42 crc kubenswrapper[4847]: E1210 14:24:42.759451 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.759617 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:42 crc kubenswrapper[4847]: E1210 14:24:42.759705 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.794371 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.794433 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.794449 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.794471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.794489 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.897357 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.897405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.897421 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.897444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.897461 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:42Z","lastTransitionTime":"2025-12-10T14:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.960250 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/1.log" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.965843 4847 scope.go:117] "RemoveContainer" containerID="004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565" Dec 10 14:24:42 crc kubenswrapper[4847]: E1210 14:24:42.966173 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:24:42 crc kubenswrapper[4847]: I1210 14:24:42.985729 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:42Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:42.999985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.000050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.000074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.000104 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.000130 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.007310 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.029499 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.050279 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.066295 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.092352 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.102563 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.102619 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.102632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.102652 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.102668 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.105551 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.119278 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.131585 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.145262 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.160340 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.176410 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.199797 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.205866 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.205901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.205910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.205925 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.205934 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.215217 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:43Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.309590 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.309951 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.309971 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.309995 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.310010 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.413252 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.413310 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.413325 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.413343 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.413356 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.516597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.516895 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.517023 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.517152 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.517243 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.619296 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.619621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.619759 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.619869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.619968 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.722636 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.722685 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.722694 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.722708 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.722746 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.825952 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.826243 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.826256 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.826270 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.826310 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.928283 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.928312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.928321 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.928337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:43 crc kubenswrapper[4847]: I1210 14:24:43.928347 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:43Z","lastTransitionTime":"2025-12-10T14:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.032802 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.032842 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.032853 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.032867 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.032878 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.135332 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.135400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.135412 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.135428 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.135439 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.237881 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.237927 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.237936 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.237948 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.237957 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.340253 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.340303 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.340312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.340340 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.340351 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.442835 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.442892 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.442905 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.442922 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.442933 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.531567 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.531800 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:25:00.531774121 +0000 UTC m=+50.100991761 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.545497 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.545552 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.545565 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.545582 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.545977 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.553974 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw"] Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.554510 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.556545 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.556678 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.568759 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.584973 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.599193 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.615334 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.627161 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632474 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrpk\" (UniqueName: \"kubernetes.io/projected/8a1b619f-c51f-47df-a401-4cb5cc749690-kube-api-access-dzrpk\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632505 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632537 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632566 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632595 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8a1b619f-c51f-47df-a401-4cb5cc749690-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632656 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632728 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632668 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632733 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632787 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632800 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632742 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:00.632726368 +0000 UTC m=+50.201943998 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632840 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.632866 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632941 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632955 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632964 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632971 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:00.632952654 +0000 UTC m=+50.202170284 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.632992 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:00.632983725 +0000 UTC m=+50.202201435 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.633007 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:00.633000435 +0000 UTC m=+50.202218165 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.640079 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.648255 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.648295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.648304 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.648317 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.648327 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.652695 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.663941 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.674481 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.689389 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.704507 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.715643 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.725421 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.733556 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8a1b619f-c51f-47df-a401-4cb5cc749690-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.733607 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.733638 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrpk\" (UniqueName: \"kubernetes.io/projected/8a1b619f-c51f-47df-a401-4cb5cc749690-kube-api-access-dzrpk\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.733656 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.734292 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.734495 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8a1b619f-c51f-47df-a401-4cb5cc749690-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.736569 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.738748 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8a1b619f-c51f-47df-a401-4cb5cc749690-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.749791 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrpk\" (UniqueName: \"kubernetes.io/projected/8a1b619f-c51f-47df-a401-4cb5cc749690-kube-api-access-dzrpk\") pod \"ovnkube-control-plane-749d76644c-tlzxw\" (UID: \"8a1b619f-c51f-47df-a401-4cb5cc749690\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.750677 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.750734 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.750746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.750764 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.750778 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.751624 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:44Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.758850 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.758861 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.758982 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.758862 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.759120 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:44 crc kubenswrapper[4847]: E1210 14:24:44.759049 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.853298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.853338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.853347 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.853366 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.853376 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.867840 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" Dec 10 14:24:44 crc kubenswrapper[4847]: W1210 14:24:44.881005 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a1b619f_c51f_47df_a401_4cb5cc749690.slice/crio-142eac8f410b544fb3d8a25e3d43eb6ae4f421872f3219c0b7ad27bc83c1d136 WatchSource:0}: Error finding container 142eac8f410b544fb3d8a25e3d43eb6ae4f421872f3219c0b7ad27bc83c1d136: Status 404 returned error can't find the container with id 142eac8f410b544fb3d8a25e3d43eb6ae4f421872f3219c0b7ad27bc83c1d136 Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.956337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.956368 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.956378 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.956392 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.956400 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:44Z","lastTransitionTime":"2025-12-10T14:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:44 crc kubenswrapper[4847]: I1210 14:24:44.974224 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" event={"ID":"8a1b619f-c51f-47df-a401-4cb5cc749690","Type":"ContainerStarted","Data":"142eac8f410b544fb3d8a25e3d43eb6ae4f421872f3219c0b7ad27bc83c1d136"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.058580 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.058619 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.058634 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.058660 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.058672 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.161549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.161585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.161594 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.161611 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.161621 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.264266 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.264326 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.264339 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.264355 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.264364 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.368676 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.368813 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.368828 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.368893 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.368908 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.471203 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.471247 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.471257 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.471272 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.471283 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.573605 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.573649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.573658 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.573672 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.573681 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.622337 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2khbq"] Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.622749 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: E1210 14:24:45.622805 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.635837 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.644406 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.644531 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnc9p\" (UniqueName: \"kubernetes.io/projected/e4db231f-2f4a-40e0-8137-c45b3f604fb6-kube-api-access-dnc9p\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.655702 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.675753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.675814 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.675825 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.675851 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.675883 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.678013 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.690908 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.701508 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.712494 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.731112 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.743424 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.745211 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.745305 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnc9p\" (UniqueName: \"kubernetes.io/projected/e4db231f-2f4a-40e0-8137-c45b3f604fb6-kube-api-access-dnc9p\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: E1210 14:24:45.745434 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:45 crc kubenswrapper[4847]: E1210 14:24:45.745541 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:46.245517323 +0000 UTC m=+35.814734953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.753818 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.762976 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnc9p\" (UniqueName: \"kubernetes.io/projected/e4db231f-2f4a-40e0-8137-c45b3f604fb6-kube-api-access-dnc9p\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.769321 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.778774 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.778833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.778843 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.778857 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.778869 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.784782 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.797917 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.813079 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.828912 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.860215 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.877278 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:45Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.881669 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.881720 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.881732 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.881746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.881756 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.979398 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" event={"ID":"8a1b619f-c51f-47df-a401-4cb5cc749690","Type":"ContainerStarted","Data":"9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231"} Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.984630 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.984698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.984736 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.984766 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:45 crc kubenswrapper[4847]: I1210 14:24:45.984781 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:45Z","lastTransitionTime":"2025-12-10T14:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.087575 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.087603 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.087611 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.087625 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.087634 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.189807 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.189843 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.189852 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.189864 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.189874 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.265061 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:46 crc kubenswrapper[4847]: E1210 14:24:46.265224 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:46 crc kubenswrapper[4847]: E1210 14:24:46.265289 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:47.265272586 +0000 UTC m=+36.834490216 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.292776 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.292830 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.292843 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.292861 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.292873 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.395768 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.395803 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.395811 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.395826 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.395837 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.498355 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.498402 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.498412 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.498429 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.498442 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.600786 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.600833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.600848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.600869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.600886 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.704188 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.704461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.704469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.704482 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.704494 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.758912 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.758865 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:46 crc kubenswrapper[4847]: E1210 14:24:46.759087 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.758957 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:46 crc kubenswrapper[4847]: E1210 14:24:46.759190 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:46 crc kubenswrapper[4847]: E1210 14:24:46.759381 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.806473 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.806515 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.806527 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.806544 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.806557 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.909499 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.909535 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.909543 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.909557 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.909567 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:46Z","lastTransitionTime":"2025-12-10T14:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.984149 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" event={"ID":"8a1b619f-c51f-47df-a401-4cb5cc749690","Type":"ContainerStarted","Data":"9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4"} Dec 10 14:24:46 crc kubenswrapper[4847]: I1210 14:24:46.997343 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:46Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.005664 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.011998 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.012029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.012038 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.012050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.012059 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.017624 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.029544 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.039685 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.049510 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.062824 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.074884 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.084887 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.109067 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.113681 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.113759 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.113774 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.113790 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.113802 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.122013 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.132458 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.140666 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.151366 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.160795 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.169973 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:47Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.215999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.216043 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.216057 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.216072 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.216085 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.272813 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:47 crc kubenswrapper[4847]: E1210 14:24:47.273004 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:47 crc kubenswrapper[4847]: E1210 14:24:47.273096 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:49.273077306 +0000 UTC m=+38.842294936 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.318479 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.318527 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.318537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.318554 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.318567 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.420931 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.420976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.420988 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.421007 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.421019 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.523361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.523412 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.523422 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.523437 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.523447 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.625962 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.626015 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.626031 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.626052 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.626069 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.728251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.728308 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.728320 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.728337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.728347 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.758545 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:47 crc kubenswrapper[4847]: E1210 14:24:47.758704 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.830154 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.830211 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.830223 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.830235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.830244 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.933166 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.933208 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.933221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.933237 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:47 crc kubenswrapper[4847]: I1210 14:24:47.933248 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:47Z","lastTransitionTime":"2025-12-10T14:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.035129 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.035177 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.035186 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.035198 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.035207 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.137311 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.137370 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.137382 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.137401 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.137413 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.239435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.239477 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.239485 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.239499 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.239509 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.341409 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.341450 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.341462 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.341475 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.341483 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.443961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.444012 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.444024 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.444039 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.444049 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.546486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.546529 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.546537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.546550 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.546559 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.649597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.649631 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.649640 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.649655 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.649663 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.751157 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.751212 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.751222 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.751259 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.751272 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.758995 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.759052 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:48 crc kubenswrapper[4847]: E1210 14:24:48.759124 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:48 crc kubenswrapper[4847]: E1210 14:24:48.759293 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.758997 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:48 crc kubenswrapper[4847]: E1210 14:24:48.759466 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.853969 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.854016 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.854027 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.854044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.854056 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.957078 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.957752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.957885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.957982 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:48 crc kubenswrapper[4847]: I1210 14:24:48.958068 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:48Z","lastTransitionTime":"2025-12-10T14:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.062101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.062172 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.062194 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.062218 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.062235 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.165920 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.165990 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.166014 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.166040 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.166211 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.268664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.268729 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.268783 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.268807 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.268818 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.293391 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:49 crc kubenswrapper[4847]: E1210 14:24:49.293550 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:49 crc kubenswrapper[4847]: E1210 14:24:49.293612 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:24:53.293592728 +0000 UTC m=+42.862810358 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.370957 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.371001 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.371009 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.371023 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.371032 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.473368 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.473406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.473422 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.473437 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.473447 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.576120 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.576170 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.576181 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.576202 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.576215 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.678999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.679032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.679041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.679053 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.679062 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.758816 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:49 crc kubenswrapper[4847]: E1210 14:24:49.758944 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.781958 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.782013 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.782024 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.782041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.782053 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.884401 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.884484 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.884514 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.884542 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.884569 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.988300 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.988347 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.988361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.988380 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:49 crc kubenswrapper[4847]: I1210 14:24:49.988392 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:49Z","lastTransitionTime":"2025-12-10T14:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.091133 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.091173 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.091185 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.091204 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.091218 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.193651 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.193737 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.193756 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.193783 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.193802 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.296667 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.296741 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.296753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.296770 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.296782 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.398836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.398888 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.398900 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.398916 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.398928 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.500824 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.500884 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.500901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.500918 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.500929 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.603957 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.604214 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.604313 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.604402 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.604484 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.708011 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.708086 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.708097 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.708117 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.708134 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.758743 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:50 crc kubenswrapper[4847]: E1210 14:24:50.759025 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.758881 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.758858 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:50 crc kubenswrapper[4847]: E1210 14:24:50.759455 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:50 crc kubenswrapper[4847]: E1210 14:24:50.759519 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.772829 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.790969 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.804023 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.810284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.810341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.810364 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.810390 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.810409 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.814739 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.824904 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.835612 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.850001 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.866211 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.876633 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.891229 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.902976 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.912549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.912579 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.912591 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.912607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.912619 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:50Z","lastTransitionTime":"2025-12-10T14:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:50 crc kubenswrapper[4847]: I1210 14:24:50.915656 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.015041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.015090 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.015103 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.015120 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.015132 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.118435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.118500 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.118512 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.118532 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.118546 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.220993 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.221032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.221044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.221060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.221069 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.323314 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.323373 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.323386 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.323404 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.323425 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.398964 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:50Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.413886 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.426300 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.426346 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.426358 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.426374 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.426385 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.427564 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.439812 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.529011 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.529067 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.529077 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.529093 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.529103 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.631858 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.631907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.631916 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.631930 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.631940 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.700659 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.700688 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.700696 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.700728 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.700741 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.714013 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.717489 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.717537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.717549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.717566 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.717578 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.730623 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.733768 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.733813 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.733826 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.733847 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.733859 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.744357 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.747568 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.747597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.747607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.747622 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.747632 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.758944 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.759102 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.760120 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.763348 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.763383 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.763392 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.763406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.763415 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.774820 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:51Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:51 crc kubenswrapper[4847]: E1210 14:24:51.774942 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.776487 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.776531 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.776540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.776555 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.776568 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.878791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.878831 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.878843 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.878860 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.878872 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.981389 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.981438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.981448 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.981461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:51 crc kubenswrapper[4847]: I1210 14:24:51.981471 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:51Z","lastTransitionTime":"2025-12-10T14:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.083944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.084484 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.084605 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.084753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.084848 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.187354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.187392 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.187403 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.187417 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.187428 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.290090 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.290130 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.290139 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.290152 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.290162 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.392649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.392699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.392733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.392752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.392763 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.494987 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.495028 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.495040 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.495056 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.495068 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.597371 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.597632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.597695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.597791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.597866 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.700669 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.700965 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.701138 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.701266 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.701363 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.758733 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.758789 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:52 crc kubenswrapper[4847]: E1210 14:24:52.759146 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.758830 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:52 crc kubenswrapper[4847]: E1210 14:24:52.759526 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:52 crc kubenswrapper[4847]: E1210 14:24:52.759253 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.803653 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.803693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.803703 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.803761 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.803773 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.906116 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.906156 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.906164 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.906177 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:52 crc kubenswrapper[4847]: I1210 14:24:52.906185 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:52Z","lastTransitionTime":"2025-12-10T14:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.008398 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.008441 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.008456 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.008477 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.008494 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.110581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.110621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.110634 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.110649 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.110667 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.214269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.214339 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.214358 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.214377 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.214390 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.317020 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.317075 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.317085 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.317099 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.317109 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.332598 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:53 crc kubenswrapper[4847]: E1210 14:24:53.332837 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:53 crc kubenswrapper[4847]: E1210 14:24:53.332895 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:01.332879224 +0000 UTC m=+50.902096854 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.419907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.419942 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.419951 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.419963 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.419972 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.522403 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.522451 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.522461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.522476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.522485 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.625162 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.625262 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.625282 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.625306 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.625324 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.727491 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.727539 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.727549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.727560 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.727570 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.759193 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:53 crc kubenswrapper[4847]: E1210 14:24:53.759364 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.830269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.830320 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.830337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.830360 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.830379 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.933221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.933249 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.933257 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.933270 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:53 crc kubenswrapper[4847]: I1210 14:24:53.933279 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:53Z","lastTransitionTime":"2025-12-10T14:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.035888 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.035954 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.035961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.035975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.035986 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.138800 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.138852 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.138864 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.138881 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.138894 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.241362 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.241398 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.241406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.241420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.241429 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.345155 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.345195 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.345204 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.345220 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.345231 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.448068 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.448115 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.448127 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.448145 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.448157 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.550555 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.550597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.550609 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.550624 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.550636 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.653386 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.653469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.653504 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.653537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.653565 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.758493 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.758565 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.758610 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:54 crc kubenswrapper[4847]: E1210 14:24:54.758659 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:54 crc kubenswrapper[4847]: E1210 14:24:54.758771 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:54 crc kubenswrapper[4847]: E1210 14:24:54.758870 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.759554 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.759573 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.759582 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.759596 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.759607 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.861132 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.861167 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.861176 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.861189 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.861198 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.962793 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.962842 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.962853 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.962869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:54 crc kubenswrapper[4847]: I1210 14:24:54.962880 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:54Z","lastTransitionTime":"2025-12-10T14:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.065654 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.065779 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.065806 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.065838 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.065863 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.168233 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.168265 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.168273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.168284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.168293 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.271227 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.271290 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.271307 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.271331 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.271347 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.374254 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.374303 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.374313 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.374330 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.374341 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.477285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.477368 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.477392 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.477420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.477443 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.579706 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.579828 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.579866 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.579897 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.579922 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.681891 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.681941 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.681951 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.681964 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.681974 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.758995 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:55 crc kubenswrapper[4847]: E1210 14:24:55.759178 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.784231 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.784271 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.784280 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.784295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.784307 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.886618 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.886650 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.886659 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.886672 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.886681 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.989515 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.989569 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.989592 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.989618 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:55 crc kubenswrapper[4847]: I1210 14:24:55.989632 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:55Z","lastTransitionTime":"2025-12-10T14:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.092353 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.092388 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.092426 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.092470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.092483 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.194944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.194982 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.194990 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.195005 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.195014 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.297424 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.297471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.297487 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.297506 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.297520 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.400265 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.400312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.400322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.400339 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.400352 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.503205 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.503244 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.503253 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.503268 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.503277 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.605595 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.605650 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.605663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.605679 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.605691 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.708317 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.708373 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.708387 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.708405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.708416 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.759252 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.759318 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.759318 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:56 crc kubenswrapper[4847]: E1210 14:24:56.760758 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:56 crc kubenswrapper[4847]: E1210 14:24:56.760968 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:56 crc kubenswrapper[4847]: E1210 14:24:56.761126 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.765437 4847 scope.go:117] "RemoveContainer" containerID="004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.810545 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.810581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.810593 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.810610 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.810621 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.913318 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.913363 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.913377 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.913398 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:56 crc kubenswrapper[4847]: I1210 14:24:56.913413 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:56Z","lastTransitionTime":"2025-12-10T14:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.015503 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.015537 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.015546 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.015560 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.015568 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.017461 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/1.log" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.022480 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.025513 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.046005 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.058188 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.071774 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.086262 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.099171 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.110283 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.117589 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.117632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.117641 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.117657 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.117666 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.121425 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.135195 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.150654 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.160487 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.179047 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.193789 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.209957 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.219517 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.219549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.219559 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.219571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.219581 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.267766 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.285332 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.297384 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.322175 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.322295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.322306 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.322341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.322350 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.424623 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.424663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.424673 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.424687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.424697 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.526684 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.526966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.526976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.526990 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.526999 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.629193 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.629270 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.629287 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.629314 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.629331 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.732082 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.732140 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.732152 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.732169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.732183 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.758704 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:57 crc kubenswrapper[4847]: E1210 14:24:57.759212 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.836279 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.836321 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.836332 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.836347 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.836356 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.939216 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.939275 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.939291 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.939310 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:57 crc kubenswrapper[4847]: I1210 14:24:57.939320 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:57Z","lastTransitionTime":"2025-12-10T14:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.027149 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/2.log" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.028011 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/1.log" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.030680 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" exitCode=1 Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.030733 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.030768 4847 scope.go:117] "RemoveContainer" containerID="004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.031305 4847 scope.go:117] "RemoveContainer" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" Dec 10 14:24:58 crc kubenswrapper[4847]: E1210 14:24:58.031461 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.042206 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.042241 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.042249 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.042263 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.042274 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.045158 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.055856 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.066461 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.079227 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.094963 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.107873 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.120041 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.132077 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.144406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.144445 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.144457 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.144473 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.144486 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.152695 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://004f509a1504e0cef4c6851f530e299a93f2af596c21e98d88718abbd501d565\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:41Z\\\",\\\"message\\\":\\\"one reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703174 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-console/downloads]} name:Service_openshift-console/downloads_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.213:80:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.703251 6229 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 14:24:41.700688 6229 o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.165109 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.174485 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.185657 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.196695 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.205785 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.215980 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.226576 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:58Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.247380 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.247423 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.247434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.247449 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.247461 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.349975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.350008 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.350017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.350032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.350040 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.453224 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.453295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.453318 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.453347 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.453370 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.556506 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.556560 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.556571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.556590 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.556602 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.660019 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.660084 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.660101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.660130 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.660142 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.759305 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.759323 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.759374 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:24:58 crc kubenswrapper[4847]: E1210 14:24:58.759484 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:24:58 crc kubenswrapper[4847]: E1210 14:24:58.759561 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:24:58 crc kubenswrapper[4847]: E1210 14:24:58.759622 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.764383 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.764421 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.764435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.764454 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.764468 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.867458 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.867535 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.867549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.867572 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.867587 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.969902 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.969973 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.969997 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.970025 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:58 crc kubenswrapper[4847]: I1210 14:24:58.970045 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:58Z","lastTransitionTime":"2025-12-10T14:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.036358 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/2.log" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.041387 4847 scope.go:117] "RemoveContainer" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" Dec 10 14:24:59 crc kubenswrapper[4847]: E1210 14:24:59.041977 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.061272 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.072450 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.072494 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.072509 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.072531 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.072546 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.084215 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.102516 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.115748 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.126906 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.138238 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.151194 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.164619 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.174597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.174633 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.174645 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.174661 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.174673 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.176502 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.186896 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.199427 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.211931 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.224160 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.234593 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.247207 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.256326 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:59Z is after 2025-08-24T17:21:41Z" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.276862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.276910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.276922 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.276934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.276942 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.378956 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.379003 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.379012 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.379024 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.379033 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.481952 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.482028 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.482052 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.482084 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.482107 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.584423 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.584518 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.584534 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.584560 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.584580 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.687908 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.687949 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.687960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.687977 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.687992 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.759225 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:24:59 crc kubenswrapper[4847]: E1210 14:24:59.759364 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.789995 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.790027 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.790035 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.790048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.790057 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.892919 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.892958 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.892970 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.892987 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.892996 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.995836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.996460 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.996475 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.996494 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:24:59 crc kubenswrapper[4847]: I1210 14:24:59.996504 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:24:59Z","lastTransitionTime":"2025-12-10T14:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.099730 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.099788 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.099805 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.099826 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.099841 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.202076 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.202177 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.202193 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.202210 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.202223 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.304638 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.304771 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.304799 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.304893 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.304919 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.408006 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.408073 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.408091 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.408113 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.408131 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.511791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.511854 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.511878 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.511907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.511922 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.613103 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.613301 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:25:32.613275603 +0000 UTC m=+82.182493263 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.614874 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.614930 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.614942 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.614957 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.614969 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.714359 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.714436 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.714484 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.714525 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714533 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714621 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:32.714601559 +0000 UTC m=+82.283819199 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714677 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714704 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714746 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714810 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714819 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:32.714798095 +0000 UTC m=+82.284015795 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714850 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714929 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714982 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.714931 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:32.714900037 +0000 UTC m=+82.284117707 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.715065 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:32.715044761 +0000 UTC m=+82.284262431 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.716933 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.716958 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.716968 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.716986 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.716997 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.759474 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.759663 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.759755 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.759902 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.759754 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:00 crc kubenswrapper[4847]: E1210 14:25:00.760095 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.771286 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.787543 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.802630 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.816493 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.819263 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.819305 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.819317 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.819335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.819347 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.830563 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.844275 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.897787 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.918751 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.921381 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.921406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.921415 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.921429 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.921438 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:00Z","lastTransitionTime":"2025-12-10T14:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.929739 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.943050 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.956303 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.966750 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.975493 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:00 crc kubenswrapper[4847]: I1210 14:25:00.990477 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.002625 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:00Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.018339 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.024244 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.024279 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.024289 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.024304 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.024315 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.115092 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.125829 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.126424 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.126470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.126485 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.126503 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.126515 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.127877 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.137510 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.150947 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.162820 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.173267 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.183798 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.193908 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.202660 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.218160 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.228506 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.228550 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.228562 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.228577 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.228588 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.230254 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.239357 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.252310 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.267254 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.279519 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.292533 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.306907 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:01Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.331338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.331386 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.331407 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.331430 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.331445 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.421302 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:01 crc kubenswrapper[4847]: E1210 14:25:01.421515 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:01 crc kubenswrapper[4847]: E1210 14:25:01.421638 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:17.421613335 +0000 UTC m=+66.990831145 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.434262 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.434333 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.434348 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.434372 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.434390 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.536972 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.537019 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.537028 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.537043 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.537056 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.639498 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.639542 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.639554 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.639570 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.639584 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.742425 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.742500 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.742524 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.742547 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.742562 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.759140 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:01 crc kubenswrapper[4847]: E1210 14:25:01.759343 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.845671 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.845736 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.845750 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.845767 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.845779 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.949506 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.949547 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.949555 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.949571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:01 crc kubenswrapper[4847]: I1210 14:25:01.949582 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:01Z","lastTransitionTime":"2025-12-10T14:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.038581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.038664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.038690 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.038759 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.038780 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.055016 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:02Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.064405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.064455 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.064465 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.064483 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.064496 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.082280 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:02Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.086848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.086890 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.086901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.086917 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.086927 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.105358 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:02Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.110095 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.110139 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.110150 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.110169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.110182 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.127512 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:02Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.131727 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.131783 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.131802 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.131822 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.131841 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.149926 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:02Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.150058 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.151752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.151806 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.151820 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.151837 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.151850 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.254001 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.254045 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.254058 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.254074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.254085 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.356338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.356388 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.356400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.356416 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.356428 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.458766 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.458797 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.458805 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.458819 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.458831 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.561381 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.561450 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.561472 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.561503 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.561524 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.665002 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.665062 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.665076 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.665093 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.665104 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.759304 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.759304 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.759307 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.759661 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.759742 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:02 crc kubenswrapper[4847]: E1210 14:25:02.759466 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.767224 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.767272 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.767284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.767299 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.767311 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.868963 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.868998 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.869026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.869044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.869055 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.971589 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.971665 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.971675 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.971693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:02 crc kubenswrapper[4847]: I1210 14:25:02.971704 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:02Z","lastTransitionTime":"2025-12-10T14:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.074130 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.074172 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.074186 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.074202 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.074214 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.176039 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.176097 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.176109 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.176124 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.176133 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.279335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.279393 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.279409 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.279432 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.279449 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.382017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.382050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.382058 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.382070 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.382077 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.484548 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.484642 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.484667 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.484695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.484753 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.587285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.587332 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.587352 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.587375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.587392 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.690215 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.690255 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.690276 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.690292 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.690303 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.759446 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:03 crc kubenswrapper[4847]: E1210 14:25:03.759609 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.793096 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.793133 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.793143 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.793158 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.793167 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.895477 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.895531 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.895541 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.895556 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.895568 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.998651 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.998699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.998729 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.998747 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:03 crc kubenswrapper[4847]: I1210 14:25:03.998759 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:03Z","lastTransitionTime":"2025-12-10T14:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.102058 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.102104 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.102121 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.102142 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.102159 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.204937 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.205033 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.205050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.205074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.205091 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.308539 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.308682 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.308764 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.308795 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.308816 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.412324 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.412397 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.412420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.412447 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.412470 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.515935 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.516041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.516077 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.516105 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.516125 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.618962 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.619074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.619092 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.619110 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.619122 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.722147 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.722226 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.722250 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.722281 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.722307 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.759043 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.759080 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.759201 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:04 crc kubenswrapper[4847]: E1210 14:25:04.759343 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:04 crc kubenswrapper[4847]: E1210 14:25:04.759562 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:04 crc kubenswrapper[4847]: E1210 14:25:04.759710 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.824208 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.824244 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.824252 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.824264 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.824273 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.926230 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.926267 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.926278 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.926291 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:04 crc kubenswrapper[4847]: I1210 14:25:04.926300 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:04Z","lastTransitionTime":"2025-12-10T14:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.028423 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.028458 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.028471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.028487 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.028499 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.131219 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.131259 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.131269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.131285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.131295 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.234021 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.234075 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.234085 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.234100 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.234111 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.336538 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.336572 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.336580 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.336592 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.336601 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.439322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.439410 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.439447 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.439486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.439508 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.542206 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.542257 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.542268 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.542286 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.542298 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.644420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.644477 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.644492 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.644508 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.644521 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.746925 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.746959 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.746968 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.746981 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.746991 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.759441 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:05 crc kubenswrapper[4847]: E1210 14:25:05.759583 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.849611 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.849660 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.849670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.849687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.849700 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.951674 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.951741 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.951758 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.951779 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:05 crc kubenswrapper[4847]: I1210 14:25:05.951792 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:05Z","lastTransitionTime":"2025-12-10T14:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.054907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.054960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.054983 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.055005 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.055022 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.158437 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.158528 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.158568 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.158600 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.158623 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.262087 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.262149 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.262170 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.262192 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.262206 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.364797 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.364847 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.364888 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.364913 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.364928 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.468388 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.468461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.468483 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.468511 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.468532 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.571636 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.571784 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.571821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.571849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.571911 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.674977 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.675025 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.675034 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.675048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.675057 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.759346 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.759402 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:06 crc kubenswrapper[4847]: E1210 14:25:06.759482 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.759552 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:06 crc kubenswrapper[4847]: E1210 14:25:06.759586 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:06 crc kubenswrapper[4847]: E1210 14:25:06.759698 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.777532 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.777594 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.777618 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.777641 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.777654 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.881251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.881311 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.881322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.881338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.881349 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.984116 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.984156 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.984165 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.984178 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:06 crc kubenswrapper[4847]: I1210 14:25:06.984187 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:06Z","lastTransitionTime":"2025-12-10T14:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.086774 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.086822 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.086834 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.086849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.086859 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.189195 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.189257 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.189268 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.189288 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.189303 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.291927 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.291966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.291976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.292013 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.292025 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.394836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.395159 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.395170 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.395185 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.395197 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.496788 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.496826 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.496834 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.496847 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.496857 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.599246 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.599284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.599292 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.599305 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.599314 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.701424 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.701466 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.701484 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.701498 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.701507 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.759183 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:07 crc kubenswrapper[4847]: E1210 14:25:07.759334 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.803307 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.803341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.803350 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.803363 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.803372 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.906112 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.906161 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.906175 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.906192 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:07 crc kubenswrapper[4847]: I1210 14:25:07.906205 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:07Z","lastTransitionTime":"2025-12-10T14:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.009817 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.009867 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.009876 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.009890 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.009899 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.112877 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.112916 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.112925 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.112943 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.112953 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.215617 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.215663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.215674 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.215689 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.215699 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.317800 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.317833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.317845 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.317866 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.317881 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.420617 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.420692 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.420755 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.420785 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.420807 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.523482 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.523526 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.523536 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.523550 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.523558 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.625262 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.625303 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.625313 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.625330 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.625340 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.727804 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.727865 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.727920 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.727943 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.727963 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.759005 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.759050 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:08 crc kubenswrapper[4847]: E1210 14:25:08.759217 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.759282 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:08 crc kubenswrapper[4847]: E1210 14:25:08.759396 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:08 crc kubenswrapper[4847]: E1210 14:25:08.759535 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.830603 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.830675 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.830688 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.830733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.830749 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.933632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.933753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.933775 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.933798 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:08 crc kubenswrapper[4847]: I1210 14:25:08.933816 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:08Z","lastTransitionTime":"2025-12-10T14:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.036876 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.036955 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.036976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.037002 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.037023 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.140029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.140073 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.140085 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.140101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.140112 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.243630 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.243698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.243752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.243774 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.243794 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.346772 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.346833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.346848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.346869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.346886 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.449456 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.449492 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.449502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.449518 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.449530 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.552877 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.552935 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.552948 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.552982 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.552998 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.656192 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.656232 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.656244 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.656259 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.656273 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759088 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:09 crc kubenswrapper[4847]: E1210 14:25:09.759277 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759517 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759551 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759571 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.759584 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.861175 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.861204 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.861214 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.861235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.861250 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.963500 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.963540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.963552 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.963567 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:09 crc kubenswrapper[4847]: I1210 14:25:09.963579 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:09Z","lastTransitionTime":"2025-12-10T14:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.066180 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.066240 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.066260 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.066280 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.066294 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.168950 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.168982 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.168992 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.169007 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.169018 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.271668 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.271738 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.271748 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.271762 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.271773 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.375502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.375559 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.375586 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.375614 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.375668 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.478453 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.478562 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.478581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.478607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.478628 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.581784 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.581853 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.581870 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.581895 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.581912 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.684824 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.684878 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.684895 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.684917 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.684934 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.758801 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.758814 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:10 crc kubenswrapper[4847]: E1210 14:25:10.759132 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:10 crc kubenswrapper[4847]: E1210 14:25:10.759177 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.758848 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:10 crc kubenswrapper[4847]: E1210 14:25:10.759270 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.773340 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.788541 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.788594 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.788606 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.788622 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.788635 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.791353 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.805907 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.819924 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.831216 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.842967 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.870542 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.882669 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.891849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.891897 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.891909 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.891927 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.891937 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.898799 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.914787 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.928943 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.939381 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.951394 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.965684 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.981187 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.994523 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.994590 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.994599 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.994613 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.994640 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:10Z","lastTransitionTime":"2025-12-10T14:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:10 crc kubenswrapper[4847]: I1210 14:25:10.996091 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:10Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.006772 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:11Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.098184 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.098238 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.098248 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.098264 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.098274 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.201944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.201989 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.202006 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.202032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.202043 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.306050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.306098 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.306110 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.306126 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.306137 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.408752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.408814 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.408830 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.408856 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.408882 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.510961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.511003 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.511013 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.511030 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.511042 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.612831 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.613458 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.613528 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.613621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.613689 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.715670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.715729 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.715757 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.715778 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.715794 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.758690 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:11 crc kubenswrapper[4847]: E1210 14:25:11.758842 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.818517 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.818573 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.818585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.818603 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.818615 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.920753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.920794 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.920804 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.920821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:11 crc kubenswrapper[4847]: I1210 14:25:11.920832 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:11Z","lastTransitionTime":"2025-12-10T14:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.024585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.024625 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.024639 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.024657 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.024668 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.127685 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.127755 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.127769 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.127790 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.127804 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.230297 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.230333 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.230362 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.230378 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.230388 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.331942 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.331975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.332023 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.332037 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.332047 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.434523 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.434585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.434597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.434611 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.434623 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.449250 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.449281 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.449289 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.449304 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.449312 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.463415 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:12Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.467413 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.467454 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.467466 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.467484 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.467495 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.478008 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:12Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.481213 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.481262 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.481270 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.481284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.481292 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.494435 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:12Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.499642 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.499689 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.499697 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.499728 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.499741 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.513589 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:12Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.517971 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.518018 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.518032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.518054 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.518066 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.530040 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:12Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.530157 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.537094 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.537139 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.537152 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.537171 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.537183 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.640005 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.640039 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.640048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.640063 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.640075 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.742334 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.742363 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.742372 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.742387 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.742412 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.758963 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.759019 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.759085 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.759114 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.759161 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:12 crc kubenswrapper[4847]: E1210 14:25:12.759283 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.845270 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.845331 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.845346 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.845361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.845372 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.947845 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.947877 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.947886 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.947899 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:12 crc kubenswrapper[4847]: I1210 14:25:12.947908 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:12Z","lastTransitionTime":"2025-12-10T14:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.050218 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.050253 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.050264 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.050277 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.050287 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.155416 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.155528 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.155544 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.155566 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.155589 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.259106 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.259179 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.259192 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.259211 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.259251 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.362320 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.362372 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.362384 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.362400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.362410 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.465890 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.465951 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.465964 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.465980 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.465991 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.567648 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.567687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.567698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.567734 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.567748 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.670546 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.670590 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.670602 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.670621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.670633 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.759274 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:13 crc kubenswrapper[4847]: E1210 14:25:13.759504 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.760388 4847 scope.go:117] "RemoveContainer" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" Dec 10 14:25:13 crc kubenswrapper[4847]: E1210 14:25:13.760629 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.772769 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.772819 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.772832 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.772849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.772861 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.875789 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.875885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.875901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.875926 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.875943 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.978540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.978574 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.978583 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.978597 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:13 crc kubenswrapper[4847]: I1210 14:25:13.978606 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:13Z","lastTransitionTime":"2025-12-10T14:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.080593 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.080662 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.080676 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.080693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.080727 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.183235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.183298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.183312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.183327 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.183338 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.286664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.286770 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.286785 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.286805 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.286860 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.389475 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.389533 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.389551 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.389573 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.389590 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.492218 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.492258 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.492269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.492285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.492296 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.594426 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.594456 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.594464 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.594476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.594488 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.697836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.697894 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.697909 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.697928 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.697944 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.758979 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.759002 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.759030 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:14 crc kubenswrapper[4847]: E1210 14:25:14.759100 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:14 crc kubenswrapper[4847]: E1210 14:25:14.759200 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:14 crc kubenswrapper[4847]: E1210 14:25:14.759329 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.799840 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.799900 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.799917 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.799943 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.799963 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.901650 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.901688 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.901699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.901728 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:14 crc kubenswrapper[4847]: I1210 14:25:14.901757 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:14Z","lastTransitionTime":"2025-12-10T14:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.004073 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.004118 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.004128 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.004144 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.004153 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.106892 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.106947 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.106957 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.106972 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.106983 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.209638 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.209699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.209748 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.209767 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.209781 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.312316 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.312355 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.312364 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.312396 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.312410 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.415769 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.415812 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.415821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.415835 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.415847 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.517821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.517862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.517871 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.517885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.517893 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.619887 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.619925 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.619934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.619948 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.619957 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.722648 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.722731 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.722746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.722768 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.722781 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.759124 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:15 crc kubenswrapper[4847]: E1210 14:25:15.759282 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.824895 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.824979 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.824992 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.825033 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.825052 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.929729 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.929808 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.929819 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.929834 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:15 crc kubenswrapper[4847]: I1210 14:25:15.929848 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:15Z","lastTransitionTime":"2025-12-10T14:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.032362 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.032394 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.032403 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.032416 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.032424 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.135125 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.135171 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.135181 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.135196 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.135207 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.238022 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.238061 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.238070 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.238087 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.238097 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.340162 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.340204 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.340214 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.340230 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.340241 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.443189 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.443242 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.443252 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.443267 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.443278 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.546220 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.546287 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.546297 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.546311 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.546321 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.647965 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.647999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.648007 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.648021 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.648030 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.750791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.750859 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.750870 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.750883 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.750892 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.759151 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.759233 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:16 crc kubenswrapper[4847]: E1210 14:25:16.759263 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:16 crc kubenswrapper[4847]: E1210 14:25:16.759446 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.759603 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:16 crc kubenswrapper[4847]: E1210 14:25:16.759676 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.853419 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.853784 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.853952 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.854117 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.854248 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.957023 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.957063 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.957075 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.957090 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:16 crc kubenswrapper[4847]: I1210 14:25:16.957099 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:16Z","lastTransitionTime":"2025-12-10T14:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.059819 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.059857 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.059869 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.059887 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.059899 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.162979 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.163033 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.163043 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.163057 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.163066 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.265670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.265741 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.265756 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.265775 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.265788 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.368659 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.368703 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.368734 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.368749 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.368762 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.470625 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.470662 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.470670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.470684 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.470694 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.482093 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:17 crc kubenswrapper[4847]: E1210 14:25:17.482265 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:17 crc kubenswrapper[4847]: E1210 14:25:17.482363 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:25:49.482342259 +0000 UTC m=+99.051559949 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.572750 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.572818 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.572833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.572848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.572859 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.675021 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.675082 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.675096 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.675111 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.675122 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.758983 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:17 crc kubenswrapper[4847]: E1210 14:25:17.759134 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.777047 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.777094 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.777105 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.777120 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.777132 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.879915 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.879960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.879970 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.879985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.879995 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.982298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.982336 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.982349 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.982365 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:17 crc kubenswrapper[4847]: I1210 14:25:17.982376 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:17Z","lastTransitionTime":"2025-12-10T14:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.084635 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.084695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.084741 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.084762 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.084777 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.187078 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.187127 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.187138 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.187154 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.187165 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.290213 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.290254 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.290265 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.290280 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.290291 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.392784 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.392827 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.392838 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.392858 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.392871 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.494664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.494704 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.494734 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.494750 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.494763 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.596397 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.596642 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.596746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.596848 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.596960 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.698928 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.699143 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.699235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.699298 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.699354 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.758862 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.758880 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:18 crc kubenswrapper[4847]: E1210 14:25:18.759472 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:18 crc kubenswrapper[4847]: E1210 14:25:18.759317 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.758929 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:18 crc kubenswrapper[4847]: E1210 14:25:18.759671 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.801217 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.801252 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.801260 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.801274 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.801283 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.903505 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.903847 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.903932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.904045 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:18 crc kubenswrapper[4847]: I1210 14:25:18.904117 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:18Z","lastTransitionTime":"2025-12-10T14:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.006877 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.006915 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.006926 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.006941 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.006949 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.100119 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/0.log" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.100209 4847 generic.go:334] "Generic (PLEG): container finished" podID="71681159-7da8-4bc9-837c-d0e3b7397e2e" containerID="b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15" exitCode=1 Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.100261 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerDied","Data":"b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.100965 4847 scope.go:117] "RemoveContainer" containerID="b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.109063 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.109103 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.109112 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.109128 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.109146 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.119034 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.129975 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.142480 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.155886 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.170401 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.182944 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.195868 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.210039 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.211525 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.211554 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.211563 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.211578 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.211587 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.220880 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.238686 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.257119 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.270445 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.284432 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.298919 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.312619 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.314133 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.314171 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.314182 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.314199 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.314215 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.326627 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.339117 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:19Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.416017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.416047 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.416055 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.416079 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.416091 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.518016 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.518045 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.518055 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.518068 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.518077 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.621085 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.621128 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.621138 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.621167 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.621177 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.724293 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.724334 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.724343 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.724357 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.724367 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.759045 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:19 crc kubenswrapper[4847]: E1210 14:25:19.759192 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.826175 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.826212 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.826221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.826234 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.826242 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.928013 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.928053 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.928062 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.928076 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:19 crc kubenswrapper[4847]: I1210 14:25:19.928087 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:19Z","lastTransitionTime":"2025-12-10T14:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.030025 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.030072 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.030082 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.030097 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.030107 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.105316 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/0.log" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.105381 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerStarted","Data":"c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.122431 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.132103 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.132161 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.132171 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.132186 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.132196 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.134810 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.151980 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.169657 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.184443 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.195611 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.217902 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.229460 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.234342 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.234375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.234385 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.234400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.234410 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.240119 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.252038 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.270837 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.284914 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.296569 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.310520 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.324090 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.336407 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.336478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.336492 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.336513 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.336526 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.337240 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.349445 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.438434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.438470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.438479 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.438492 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.438502 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.540511 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.540543 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.540555 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.540570 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.540580 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.643586 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.643637 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.643654 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.643675 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.643690 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.747740 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.747829 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.747847 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.747881 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.747899 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.758833 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.759076 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:20 crc kubenswrapper[4847]: E1210 14:25:20.759178 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:20 crc kubenswrapper[4847]: E1210 14:25:20.759277 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.759289 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:20 crc kubenswrapper[4847]: E1210 14:25:20.759595 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.772119 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.771184 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.793644 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.806443 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.820244 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.831099 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.844352 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.850944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.850982 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.851042 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.851061 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.851074 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.861110 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.871498 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.885916 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.896179 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.907956 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.916804 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.926771 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.935758 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.950876 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.953363 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.953421 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.953434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.953450 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.953484 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:20Z","lastTransitionTime":"2025-12-10T14:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.964511 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:20 crc kubenswrapper[4847]: I1210 14:25:20.976685 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:20Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.056419 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.056476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.056498 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.056529 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.056546 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.159337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.159410 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.159426 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.159444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.159455 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.261926 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.262190 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.262296 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.262378 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.262442 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.365047 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.365086 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.365096 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.365111 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.365122 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.467683 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.467800 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.467817 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.467842 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.467857 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.569919 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.569961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.569975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.569994 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.570008 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.672370 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.672424 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.672438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.672457 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.672468 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.759130 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:21 crc kubenswrapper[4847]: E1210 14:25:21.759278 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.775229 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.775277 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.775285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.775300 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.775311 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.877664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.877740 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.877749 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.877764 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.877773 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.980564 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.980643 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.980657 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.980682 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:21 crc kubenswrapper[4847]: I1210 14:25:21.980696 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:21Z","lastTransitionTime":"2025-12-10T14:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.084009 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.084049 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.084059 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.084074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.084083 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.185927 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.185997 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.186012 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.186029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.186040 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.288443 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.288485 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.288493 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.288507 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.288518 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.391461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.391507 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.391529 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.391545 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.391557 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.493749 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.493781 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.493789 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.493802 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.493810 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.596133 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.596185 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.596198 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.596220 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.596233 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.647280 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.647328 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.647340 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.647357 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.647369 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.659742 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:22Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.662985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.663020 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.663032 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.663049 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.663058 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.674667 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:22Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.678145 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.678176 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.678187 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.678222 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.678231 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.690625 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:22Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.694213 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.694251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.694261 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.694277 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.694288 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.705372 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:22Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.708682 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.708733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.708744 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.708760 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.708770 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.719154 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:22Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.719324 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.721394 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.721444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.721456 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.721511 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.721549 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.759379 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.759486 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.759535 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.759636 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.759387 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:22 crc kubenswrapper[4847]: E1210 14:25:22.759767 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.823646 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.823672 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.823681 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.823695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.823704 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.926618 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.926677 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.926687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.926701 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:22 crc kubenswrapper[4847]: I1210 14:25:22.926734 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:22Z","lastTransitionTime":"2025-12-10T14:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.029387 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.029433 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.029449 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.029464 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.029474 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.131812 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.131851 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.131862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.131877 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.131889 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.234405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.234438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.234448 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.234460 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.234468 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.336478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.336523 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.336543 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.336564 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.336582 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.439557 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.439602 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.439612 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.439628 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.439637 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.541291 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.541331 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.541340 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.541376 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.541387 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.643862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.643916 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.643930 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.643949 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.643961 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.747471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.747520 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.747528 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.747543 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.747555 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.759122 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:23 crc kubenswrapper[4847]: E1210 14:25:23.759316 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.849575 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.849632 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.849670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.849693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.849735 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.951851 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.951893 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.951905 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.951921 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:23 crc kubenswrapper[4847]: I1210 14:25:23.951933 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:23Z","lastTransitionTime":"2025-12-10T14:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.054215 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.054251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.054260 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.054274 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.054284 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.158327 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.158361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.158371 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.158385 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.158393 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.261234 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.261293 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.261312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.261365 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.261385 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.363490 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.363556 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.363566 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.363604 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.363618 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.467153 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.467238 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.467260 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.467287 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.467305 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.569969 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.570031 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.570048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.570074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.570109 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.672726 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.672797 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.672809 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.672823 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.672844 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.759117 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.759201 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.759144 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:24 crc kubenswrapper[4847]: E1210 14:25:24.759276 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:24 crc kubenswrapper[4847]: E1210 14:25:24.759387 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:24 crc kubenswrapper[4847]: E1210 14:25:24.759519 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.775824 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.775864 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.775875 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.775894 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.775907 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.878903 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.878934 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.878944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.878960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.878971 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.981436 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.981468 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.981476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.981488 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:24 crc kubenswrapper[4847]: I1210 14:25:24.981497 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:24Z","lastTransitionTime":"2025-12-10T14:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.083673 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.083732 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.083744 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.083759 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.083770 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.187406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.187480 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.187497 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.187520 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.187543 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.289647 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.289684 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.289695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.289728 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.289737 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.392215 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.392266 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.392277 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.392295 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.392307 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.494976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.495029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.495043 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.495062 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.495076 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.597617 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.597675 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.597686 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.597705 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.597743 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.700516 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.700559 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.700569 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.700587 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.700596 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.758508 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:25 crc kubenswrapper[4847]: E1210 14:25:25.758673 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.759512 4847 scope.go:117] "RemoveContainer" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.803821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.803866 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.803879 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.803897 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.803911 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.907076 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.907121 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.907131 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.907148 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:25 crc kubenswrapper[4847]: I1210 14:25:25.907158 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:25Z","lastTransitionTime":"2025-12-10T14:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.010379 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.010435 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.010446 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.010470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.010485 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.113734 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.113802 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.113819 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.113839 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.113851 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.217518 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.217574 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.217587 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.217606 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.217621 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.320957 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.321067 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.321095 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.321130 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.321156 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.423674 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.423924 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.423948 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.423976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.423994 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.526601 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.526938 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.526947 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.526961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.526970 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.630184 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.630227 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.630238 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.630254 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.630266 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.732538 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.732578 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.732590 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.732607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.732619 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.759203 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.759233 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.759208 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:26 crc kubenswrapper[4847]: E1210 14:25:26.759564 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:26 crc kubenswrapper[4847]: E1210 14:25:26.759680 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:26 crc kubenswrapper[4847]: E1210 14:25:26.759785 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.772560 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.835361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.835407 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.835418 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.835434 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.835448 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.937708 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.937766 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.937775 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.937791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:26 crc kubenswrapper[4847]: I1210 14:25:26.937800 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:26Z","lastTransitionTime":"2025-12-10T14:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.040021 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.040077 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.040087 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.040101 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.040112 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.127911 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/2.log" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.130599 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.131324 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.142917 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.142969 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.142980 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.142995 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.143013 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.145064 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.169620 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.178705 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.188032 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.213012 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.226378 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.242467 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.244955 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.245004 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.245019 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.245038 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.245051 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.257001 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.272749 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.307202 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.323971 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.338479 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.348019 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.348073 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.348081 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.348094 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.348102 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.354177 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.366683 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.379531 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.414902 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"611278f1-00e9-4757-90f7-a425f73e4840\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.436173 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae94959-165f-451c-b3bd-9d08c4051cc1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450098 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450141 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450153 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450179 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.450978 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.465781 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:27Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.552289 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.552539 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.552644 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.552775 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.552856 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.654788 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.654831 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.654843 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.654857 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.654866 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.757533 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.757569 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.757580 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.757596 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.757609 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.758892 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:27 crc kubenswrapper[4847]: E1210 14:25:27.759008 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.860932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.860993 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.861009 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.861041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.861058 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.963921 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.963970 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.963979 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.963999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:27 crc kubenswrapper[4847]: I1210 14:25:27.964011 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:27Z","lastTransitionTime":"2025-12-10T14:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.067327 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.067372 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.067381 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.067400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.067411 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.136340 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/3.log" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.137353 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/2.log" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.140576 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" exitCode=1 Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.140614 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.140648 4847 scope.go:117] "RemoveContainer" containerID="0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.141401 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:25:28 crc kubenswrapper[4847]: E1210 14:25:28.141619 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.162075 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae94959-165f-451c-b3bd-9d08c4051cc1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.169413 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.169466 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.169478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.169496 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.169840 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.178342 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.196370 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.209133 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.221675 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.235515 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.251168 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"611278f1-00e9-4757-90f7-a425f73e4840\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.270107 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e632e6503a01de5726622cf57b255c742f642ab4d5f93c8b774aa13580e6093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:24:57Z\\\",\\\"message\\\":\\\"es.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1210 14:24:57.647148 6445 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 14:24:57.647419 6445 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:24:57Z is after 2025-08-24T17:21:41Z]\\\\nI1210 14:24:57.647427 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:27Z\\\",\\\"message\\\":\\\"fig-daemon-gvdwq openshift-multus/multus-additional-cni-plugins-v58hl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-node-b5l5q]\\\\nI1210 14:25:27.105101 6851 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1210 14:25:27.105110 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105125 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-67b5l\\\\nI1210 14:25:27.105133 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105137 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-67b5l\\\\nF1210 14:25:27.105149 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:25:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.272235 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.272290 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.272302 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.272321 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.272338 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.285547 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.299033 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.319984 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.335578 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.348971 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.360505 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.374892 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.374941 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.374995 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.375006 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.375026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.375038 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.391087 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.409166 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.425138 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.440372 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:28Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.479777 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.479832 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.479844 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.479926 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.479939 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.582591 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.582679 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.582693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.582730 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.582741 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.685629 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.685670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.685680 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.685699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.685729 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.759155 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.759155 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.759505 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:28 crc kubenswrapper[4847]: E1210 14:25:28.759734 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:28 crc kubenswrapper[4847]: E1210 14:25:28.759846 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:28 crc kubenswrapper[4847]: E1210 14:25:28.759958 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.787653 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.787725 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.787739 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.787756 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.787766 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.890679 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.890757 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.890771 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.890792 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.890807 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.994181 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.994240 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.994254 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.994274 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:28 crc kubenswrapper[4847]: I1210 14:25:28.994285 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:28Z","lastTransitionTime":"2025-12-10T14:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.097077 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.097118 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.097130 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.097146 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.097155 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.146680 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/3.log" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.150464 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:25:29 crc kubenswrapper[4847]: E1210 14:25:29.150653 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.167191 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.180424 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.199800 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.199849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.199862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.199882 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.199902 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.208279 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.223911 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"611278f1-00e9-4757-90f7-a425f73e4840\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.245416 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae94959-165f-451c-b3bd-9d08c4051cc1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.260272 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.275867 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.289356 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.302860 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.302902 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.302912 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.302932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.302944 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.310763 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:27Z\\\",\\\"message\\\":\\\"fig-daemon-gvdwq openshift-multus/multus-additional-cni-plugins-v58hl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-node-b5l5q]\\\\nI1210 14:25:27.105101 6851 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1210 14:25:27.105110 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105125 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-67b5l\\\\nI1210 14:25:27.105133 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105137 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-67b5l\\\\nF1210 14:25:27.105149 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:25:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.319783 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.329813 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.342016 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.352809 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.365838 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.379800 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.391683 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.402116 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.405416 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.405454 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.405468 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.405486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.405499 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.413231 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.423334 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:29Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.508405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.508475 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.508486 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.508503 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.508514 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.611851 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.611918 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.611933 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.611953 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.611967 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.713928 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.713964 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.713975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.713992 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.714002 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.758975 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:29 crc kubenswrapper[4847]: E1210 14:25:29.759181 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.816858 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.816918 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.816937 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.816963 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.816980 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.919377 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.919427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.919438 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.919453 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:29 crc kubenswrapper[4847]: I1210 14:25:29.919464 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:29Z","lastTransitionTime":"2025-12-10T14:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.022399 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.022471 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.022489 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.022512 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.022528 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.126695 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.127240 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.127308 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.127380 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.127454 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.230971 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.231335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.231514 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.231668 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.231860 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.334484 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.334914 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.335050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.335242 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.335590 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.438135 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.438480 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.438762 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.438821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.438843 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.541029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.541060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.541068 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.541081 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.541089 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.643841 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.643878 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.643889 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.643903 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.643914 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.746907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.746947 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.746958 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.746975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.746986 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.758972 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.759093 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:30 crc kubenswrapper[4847]: E1210 14:25:30.759194 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.759229 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:30 crc kubenswrapper[4847]: E1210 14:25:30.759358 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:30 crc kubenswrapper[4847]: E1210 14:25:30.759426 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.774177 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.787003 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.801147 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"611278f1-00e9-4757-90f7-a425f73e4840\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.830884 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae94959-165f-451c-b3bd-9d08c4051cc1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.847532 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.849669 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.849737 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.849757 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.849780 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.849797 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.866096 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.880648 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.894618 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.912368 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:27Z\\\",\\\"message\\\":\\\"fig-daemon-gvdwq openshift-multus/multus-additional-cni-plugins-v58hl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-node-b5l5q]\\\\nI1210 14:25:27.105101 6851 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1210 14:25:27.105110 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105125 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-67b5l\\\\nI1210 14:25:27.105133 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105137 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-67b5l\\\\nF1210 14:25:27.105149 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:25:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.936569 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.952135 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.952169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.952177 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.952190 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.952198 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:30Z","lastTransitionTime":"2025-12-10T14:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.965669 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.975755 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:30 crc kubenswrapper[4847]: I1210 14:25:30.989044 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.001288 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:30Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.011973 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.022587 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.040128 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.052030 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.054837 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.054887 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.054899 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.054920 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.054934 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.063196 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:31Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.156615 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.156673 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.156685 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.156701 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.156733 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.259859 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.259925 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.259951 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.259979 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.259999 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.363335 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.363378 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.363390 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.363406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.363418 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.466359 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.466450 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.466463 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.466481 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.466490 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.569708 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.569810 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.569833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.569862 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.569885 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.672804 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.672901 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.672943 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.672976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.673031 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.758571 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:31 crc kubenswrapper[4847]: E1210 14:25:31.758838 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.775910 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.776044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.776086 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.776125 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.776150 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.879173 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.879227 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.879237 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.879254 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.879268 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.982017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.982064 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.982076 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.982094 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:31 crc kubenswrapper[4847]: I1210 14:25:31.982106 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:31Z","lastTransitionTime":"2025-12-10T14:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.084060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.084093 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.084102 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.084117 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.084128 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.186789 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.186822 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.186833 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.186868 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.186881 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.289549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.289622 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.289631 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.289645 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.289656 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.392757 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.392809 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.392820 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.392835 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.392849 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.496116 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.496168 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.496180 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.496196 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.496208 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.598785 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.598837 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.598849 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.598865 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.598876 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.659026 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.659308 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:36.659274982 +0000 UTC m=+146.228492622 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.701264 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.701308 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.701324 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.701339 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.701349 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.752236 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.752289 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.752302 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.752322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.752335 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.758686 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.758867 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.758976 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.759063 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759234 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759448 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.759528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.759584 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.759642 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.759688 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759757 4847 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759763 4847 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759827 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:36.759807408 +0000 UTC m=+146.329025038 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759861 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:36.759846639 +0000 UTC m=+146.329064279 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759865 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759898 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759917 4847 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759865 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759997 4847 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.760018 4847 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.759963 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:36.759950882 +0000 UTC m=+146.329168522 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.760110 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:36.760083856 +0000 UTC m=+146.329301526 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.772937 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.778840 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.778918 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.778932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.778954 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.778971 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.792764 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.796478 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.796522 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.796534 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.796550 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.796560 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.809202 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.813128 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.813158 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.813169 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.813186 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.813200 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.827524 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.831906 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.831950 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.831963 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.831981 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.831994 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.844992 4847 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ee37509e-2f06-438f-90b6-27523eac2f9d\\\",\\\"systemUUID\\\":\\\"7a2b8039-75e9-4d12-a71a-bbfc9770e7fe\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:32Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:32 crc kubenswrapper[4847]: E1210 14:25:32.845142 4847 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.846966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.847006 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.847018 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.847036 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.847048 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.949436 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.949490 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.949502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.949522 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:32 crc kubenswrapper[4847]: I1210 14:25:32.949535 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:32Z","lastTransitionTime":"2025-12-10T14:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.051661 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.051702 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.051738 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.051765 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.051780 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.154319 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.154375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.154389 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.154406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.154415 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.257322 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.257427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.257448 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.257476 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.257497 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.360504 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.360610 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.360655 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.360687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.360755 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.463944 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.464002 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.464018 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.464040 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.464057 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.566351 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.566664 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.566792 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.566898 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.566997 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.670342 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.670420 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.670445 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.670474 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.670490 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.759204 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:33 crc kubenswrapper[4847]: E1210 14:25:33.759437 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.772262 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.772316 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.772330 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.772348 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.772362 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.875015 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.875050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.875060 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.875074 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.875084 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.977549 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.977601 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.977613 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.977635 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:33 crc kubenswrapper[4847]: I1210 14:25:33.977648 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:33Z","lastTransitionTime":"2025-12-10T14:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.079998 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.080049 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.080065 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.080083 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.080095 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.183314 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.183405 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.183427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.183454 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.183473 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.286975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.287022 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.287034 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.287050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.287059 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.389576 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.389634 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.389645 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.389670 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.389691 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.492705 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.492762 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.492782 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.492798 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.492808 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.596400 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.596462 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.596474 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.596497 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.596510 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.698956 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.699007 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.699018 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.699036 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.699048 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.758930 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.758986 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:34 crc kubenswrapper[4847]: E1210 14:25:34.759100 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.759136 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:34 crc kubenswrapper[4847]: E1210 14:25:34.759342 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:34 crc kubenswrapper[4847]: E1210 14:25:34.759495 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.801404 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.801509 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.801521 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.801539 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.801550 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.903739 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.903781 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.903792 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.903806 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:34 crc kubenswrapper[4847]: I1210 14:25:34.903814 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:34Z","lastTransitionTime":"2025-12-10T14:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.005750 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.005798 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.005807 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.005821 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.005831 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.108669 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.108732 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.108743 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.108755 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.108764 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.211127 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.211230 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.211251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.211276 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.211297 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.313361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.313417 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.313427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.313444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.313456 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.416261 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.416329 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.416343 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.416361 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.416373 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.518907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.518956 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.518967 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.518985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.518997 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.622929 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.623241 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.623251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.623265 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.623275 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.725859 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.725903 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.725915 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.725936 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.725950 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.758529 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:35 crc kubenswrapper[4847]: E1210 14:25:35.758673 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.828752 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.828801 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.828810 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.828824 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.828833 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.931660 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.931753 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.931770 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.931790 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:35 crc kubenswrapper[4847]: I1210 14:25:35.931803 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:35Z","lastTransitionTime":"2025-12-10T14:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.034309 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.034341 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.034349 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.034362 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.034371 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.136999 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.137034 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.137042 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.137055 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.137063 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.239651 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.239697 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.239727 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.239746 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.239757 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.342667 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.342704 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.342728 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.342743 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.342753 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.445268 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.445321 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.445330 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.445351 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.445362 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.547800 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.547870 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.547892 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.547920 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.547941 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.650853 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.650932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.650956 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.650985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.651007 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.753541 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.753596 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.753607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.753621 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.753630 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.759163 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.759270 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.759179 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:36 crc kubenswrapper[4847]: E1210 14:25:36.759412 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:36 crc kubenswrapper[4847]: E1210 14:25:36.759791 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:36 crc kubenswrapper[4847]: E1210 14:25:36.759610 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.856492 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.856565 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.856603 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.856641 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.856666 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.959338 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.959406 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.959427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.959457 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:36 crc kubenswrapper[4847]: I1210 14:25:36.959482 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:36Z","lastTransitionTime":"2025-12-10T14:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.061895 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.061939 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.061949 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.061963 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.061972 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.165482 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.165576 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.165620 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.165663 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.165690 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.268337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.268368 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.268375 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.268389 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.268398 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.370909 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.370953 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.370962 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.370976 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.371002 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.473629 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.473694 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.473706 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.473748 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.473765 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.576329 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.576373 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.576412 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.576433 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.576452 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.679415 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.679447 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.679455 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.679469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.679477 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.759569 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:37 crc kubenswrapper[4847]: E1210 14:25:37.759773 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.781273 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.781304 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.781312 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.781325 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.781335 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.884048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.884097 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.884107 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.884127 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.884140 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.987211 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.987284 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.987296 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.987314 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:37 crc kubenswrapper[4847]: I1210 14:25:37.987325 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:37Z","lastTransitionTime":"2025-12-10T14:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.090299 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.090354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.090369 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.090389 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.090400 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.192938 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.193003 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.193029 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.193056 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.193074 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.295687 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.295759 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.295771 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.295791 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.295804 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.397823 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.397871 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.397885 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.397899 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.397910 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.500919 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.500964 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.500973 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.500988 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.500997 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.603469 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.603511 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.603522 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.603540 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.603550 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.705900 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.705932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.705940 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.705953 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.705962 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.758898 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.758934 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.758909 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:38 crc kubenswrapper[4847]: E1210 14:25:38.759142 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:38 crc kubenswrapper[4847]: E1210 14:25:38.759215 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:38 crc kubenswrapper[4847]: E1210 14:25:38.759308 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.808619 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.808667 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.808678 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.808693 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.808704 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.910607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.910651 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.910659 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.910676 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:38 crc kubenswrapper[4847]: I1210 14:25:38.910685 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:38Z","lastTransitionTime":"2025-12-10T14:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.013325 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.013402 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.013413 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.013427 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.013436 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.115924 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.115975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.115988 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.116012 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.116025 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.218733 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.219048 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.219134 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.219229 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.219333 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.322470 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.322538 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.322555 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.322577 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.322593 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.425585 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.425622 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.425631 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.425645 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.425655 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.527978 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.528039 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.528050 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.528069 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.528085 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.631114 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.631211 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.631227 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.631251 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.631269 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.733888 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.733960 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.733985 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.734017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.734042 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.759162 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:39 crc kubenswrapper[4847]: E1210 14:25:39.759340 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.837134 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.837199 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.837221 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.837248 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.837267 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.939932 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.939998 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.940017 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.940041 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:39 crc kubenswrapper[4847]: I1210 14:25:39.940059 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:39Z","lastTransitionTime":"2025-12-10T14:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.042961 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.043026 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.043046 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.043070 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.043090 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.146225 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.146323 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.146356 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.146374 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.146387 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.249521 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.249575 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.249583 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.249602 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.249614 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.353354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.353628 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.353672 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.353755 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.353787 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.456278 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.456326 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.456337 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.456354 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.456369 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.558623 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.558671 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.558690 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.558726 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.558743 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.661035 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.661087 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.661099 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.661115 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.661139 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.759150 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.759209 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:40 crc kubenswrapper[4847]: E1210 14:25:40.759282 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.759363 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:40 crc kubenswrapper[4847]: E1210 14:25:40.759535 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:40 crc kubenswrapper[4847]: E1210 14:25:40.759589 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.760380 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:25:40 crc kubenswrapper[4847]: E1210 14:25:40.760598 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.762936 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.762966 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.762975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.762988 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.762998 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.775392 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3f117d3a9f29f4d8f58cc5360d0b78eb41463aa2b51b689686c1e4e2fbfeaac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.786603 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5f43a20d32d21a0db2e60aa46fd059d2a9ec2cdb67efa917ad3de263e48d79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqxps\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gvdwq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.809649 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v58hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"550fdaaf-ffec-4047-b02d-3002d3fc3e2b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0baff372fbbe7b28ab604e0fe23eefde133d34de34f8e574c3de1ce7fa627de4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7400dca39d77cb6a2a1db7694ca892f1fb1802a4ad497d1ec3bd3e787052b991\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1abd171000db2ca87d64dbac3bb5f697d568c8259e5265788f070ae3f5bd6be7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3149850b41fdf2f36e4a254ac17e97f6fdb57803a1c9cefe556889351be8a2a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d89b0bf1c6830fcee88d750a10273b64a58c40f29b8b647f71b903233ad2ef07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1090828ece9cfedf6d39cc3c9329fccb48908229a87152056676c799b5ce079d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08ed56df530b3370c2d3ab016e8cf2f08084411177c5c5a0a8cb19d15403fb47\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gj4bm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v58hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.827185 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-67b5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"71681159-7da8-4bc9-837c-d0e3b7397e2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:18Z\\\",\\\"message\\\":\\\"2025-12-10T14:24:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c\\\\n2025-12-10T14:24:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_15d66fe7-4210-45f9-bd92-fb2a46ea793c to /host/opt/cni/bin/\\\\n2025-12-10T14:24:33Z [verbose] multus-daemon started\\\\n2025-12-10T14:24:33Z [verbose] Readiness Indicator file check\\\\n2025-12-10T14:25:18Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jwjkr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-multus\"/\"multus-67b5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.840607 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5x4hl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252e084b-a363-4110-95c6-57bf80cb730a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd2cf41b5341f1d45f8fe06424108f38ad61a9f3e5bcad5d5bb92c3a4a1183a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8g7jl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5x4hl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.853127 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2khbq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4db231f-2f4a-40e0-8137-c45b3f604fb6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dnc9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2khbq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.864678 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.864747 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.864758 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.864776 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.864822 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.868583 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2452c592-3665-49f6-8754-fbb109862229\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54b30a84255f4d9ff4b802b3aa060aa27a709961f1f9050718a3066492aa1c1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7290fa2431d8994e64d31677c11372e388948bd216436ea2e4a75ab1d95a2967\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1333b8c14241070d8b22c062c21fef876870d7ce262b768ae7a1db7efb47a5c7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.883149 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4afb1a8-c7a7-4eb8-bc70-a16ade5508af\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:25:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c5626e48ceee04aef4d7f5d173eecd47522d88d548549d5f6dc1aa8f6a9feba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48cef7797247c2135ed0c570e51c437d83a2ccc7028e9e703ca5ffac5c6cabb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0998b4b78e5a9138ad399f2db919e4d71c1e9fa11379746f051925d00f35d4e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a5a335338a7ccd41d9b564ca3ca7609fac9830c71dfce215c4118f412e411dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.898340 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.908990 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v9dh5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ab74b21-e521-4884-9b54-207201f6b78f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f45e6c7f88684718934c1e9a9592c79f473a44ceb8178c7632a374693c56366\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dfmr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v9dh5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.918243 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"611278f1-00e9-4757-90f7-a425f73e4840\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad26f46692f3eac3eb10485b9c2496c7f285e92f48f6eea4b09cff0c0b57ed20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6195b9f65dceeec3ad76eb4139ed12265a40f9109e562a01a8ffc093b102ebd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.939360 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ae94959-165f-451c-b3bd-9d08c4051cc1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ae02e7cf94fe6a7690394dc21af33acc8309e7801891a195d4ac443a4c8db9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3718d51bcf71616f195c88d18534260973399cfbb9471920297f76739d738ee2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bca8fd324b0d72110f648e801a3665ab8f56266f9e5c8c06e93a0fe9262edc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a835d0bd868c0f2a3e44dc4374c7ddfa7434e5991d8e19a1db06e656ee0f88ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://355bc5afa16d631ecefafacdd09a5c063e048faedcf02699c28633de5a89e11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72041b27b85c0011eb5b1bb4e6f755a7c3f04c5b76f1a92e168a3dee6f68fa45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c1661293843b627e806f9477e4e17de870f74aaab4faef2210c0ab9e4a25276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6d0af0fb8b755fa378473d362e7fdc13f50ba6db1c68dc2659efb56f1612bbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.955147 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c80a69d-2927-4296-bb72-9baec8edb31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.968444 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.968572 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.968589 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.968639 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.968656 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:40Z","lastTransitionTime":"2025-12-10T14:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.973500 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1844e1c356350312225fef9e8e7eac2639837251adc3d9dd3b8a8c6467798d18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:40 crc kubenswrapper[4847]: I1210 14:25:40.987863 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:40Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.001887 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:28Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.014545 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1b619f-c51f-47df-a401-4cb5cc749690\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cb41d35774199aba443294693fdf14eb454d05416734cbaf66354777805a231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a95e7dda8327fafd441db5e73844a9087a5d4bff5c87b2176ce92d55f740bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzrpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tlzxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.031300 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c395e06826a9522bb021729a4c4ed40a2fefb72cee30e11eefc58aa2456a0475\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43957739b6b8ee73853425c434e6f91e2a93a8e565a360b19ec0d68c76cb3ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.052167 4847 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b405b3d-ae75-409a-9e83-b098f333a5c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T14:25:27Z\\\",\\\"message\\\":\\\"fig-daemon-gvdwq openshift-multus/multus-additional-cni-plugins-v58hl openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-ovn-kubernetes/ovnkube-node-b5l5q]\\\\nI1210 14:25:27.105101 6851 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1210 14:25:27.105110 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105125 6851 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-67b5l\\\\nI1210 14:25:27.105133 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1210 14:25:27.105137 6851 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-67b5l\\\\nF1210 14:25:27.105149 6851 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T14:25:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T14:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T14:24:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T14:24:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jhhpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T14:24:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-b5l5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T14:25:41Z is after 2025-08-24T17:21:41Z" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.071365 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.071429 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.071443 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.071462 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.071476 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.173736 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.173786 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.173801 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.173822 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.173835 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.276642 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.276688 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.276699 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.276737 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.276748 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.378795 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.378827 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.378836 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.378850 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.378859 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.486031 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.486082 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.486098 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.486115 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.486127 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.589539 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.589581 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.589595 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.589614 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.589627 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.692783 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.692814 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.692822 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.692835 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.692846 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.759571 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:41 crc kubenswrapper[4847]: E1210 14:25:41.759784 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.795607 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.795661 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.795677 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.795698 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.795735 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.898357 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.898402 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.898413 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.898429 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:41 crc kubenswrapper[4847]: I1210 14:25:41.898440 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:41Z","lastTransitionTime":"2025-12-10T14:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.000841 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.000886 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.000896 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.000911 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.000922 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.103449 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.103517 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.103542 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.103568 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.103584 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.205742 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.205787 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.205796 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.205811 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.205821 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.308593 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.308629 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.308638 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.308652 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.308661 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.411127 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.411173 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.411189 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.411207 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.411217 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.513414 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.513451 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.513461 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.513474 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.513483 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.616225 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.616259 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.616269 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.616283 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.616293 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.719246 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.719285 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.719294 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.719308 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.719318 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.758949 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.758949 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.759139 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:42 crc kubenswrapper[4847]: E1210 14:25:42.759164 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:42 crc kubenswrapper[4847]: E1210 14:25:42.759304 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:42 crc kubenswrapper[4847]: E1210 14:25:42.759367 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.822422 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.822502 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.822520 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.822542 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.822556 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.925996 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.926033 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.926044 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.926059 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:42 crc kubenswrapper[4847]: I1210 14:25:42.926071 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:42Z","lastTransitionTime":"2025-12-10T14:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.029812 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.029907 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.029943 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.029975 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.030000 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:43Z","lastTransitionTime":"2025-12-10T14:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.099374 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.099437 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.099449 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.099467 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.099479 4847 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T14:25:43Z","lastTransitionTime":"2025-12-10T14:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.169234 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf"] Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.169907 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.171875 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.172053 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.172258 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.172769 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.200263 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5x4hl" podStartSLOduration=72.200243152 podStartE2EDuration="1m12.200243152s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.185786213 +0000 UTC m=+92.755003873" watchObservedRunningTime="2025-12-10 14:25:43.200243152 +0000 UTC m=+92.769460782" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.239441 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podStartSLOduration=72.239420394 podStartE2EDuration="1m12.239420394s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.238832238 +0000 UTC m=+92.808049888" watchObservedRunningTime="2025-12-10 14:25:43.239420394 +0000 UTC m=+92.808638024" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.255897 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-v58hl" podStartSLOduration=72.255878618 podStartE2EDuration="1m12.255878618s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.255270241 +0000 UTC m=+92.824487871" watchObservedRunningTime="2025-12-10 14:25:43.255878618 +0000 UTC m=+92.825096248" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.267321 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.267389 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e123a0f-959e-42ab-af61-fb62222ae80c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.267415 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e123a0f-959e-42ab-af61-fb62222ae80c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.267450 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.267564 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e123a0f-959e-42ab-af61-fb62222ae80c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.274468 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-67b5l" podStartSLOduration=72.27444237 podStartE2EDuration="1m12.27444237s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.274165013 +0000 UTC m=+92.843382663" watchObservedRunningTime="2025-12-10 14:25:43.27444237 +0000 UTC m=+92.843660000" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.294005 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.29398453 podStartE2EDuration="1m15.29398453s" podCreationTimestamp="2025-12-10 14:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.293755894 +0000 UTC m=+92.862973544" watchObservedRunningTime="2025-12-10 14:25:43.29398453 +0000 UTC m=+92.863202160" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.309233 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.309209121 podStartE2EDuration="42.309209121s" podCreationTimestamp="2025-12-10 14:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.309201421 +0000 UTC m=+92.878419051" watchObservedRunningTime="2025-12-10 14:25:43.309209121 +0000 UTC m=+92.878426751" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.334745 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-v9dh5" podStartSLOduration=72.334727405 podStartE2EDuration="1m12.334727405s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.334433067 +0000 UTC m=+92.903650717" watchObservedRunningTime="2025-12-10 14:25:43.334727405 +0000 UTC m=+92.903945035" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368790 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368830 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e123a0f-959e-42ab-af61-fb62222ae80c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368848 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e123a0f-959e-42ab-af61-fb62222ae80c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368870 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368903 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e123a0f-959e-42ab-af61-fb62222ae80c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.368943 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.369086 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8e123a0f-959e-42ab-af61-fb62222ae80c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.369738 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e123a0f-959e-42ab-af61-fb62222ae80c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.374647 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e123a0f-959e-42ab-af61-fb62222ae80c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.375237 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tlzxw" podStartSLOduration=72.375213753 podStartE2EDuration="1m12.375213753s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.374891264 +0000 UTC m=+92.944108894" watchObservedRunningTime="2025-12-10 14:25:43.375213753 +0000 UTC m=+92.944431383" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.391902 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e123a0f-959e-42ab-af61-fb62222ae80c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wk2cf\" (UID: \"8e123a0f-959e-42ab-af61-fb62222ae80c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.391950 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.391935275 podStartE2EDuration="23.391935275s" podCreationTimestamp="2025-12-10 14:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.391792941 +0000 UTC m=+92.961010591" watchObservedRunningTime="2025-12-10 14:25:43.391935275 +0000 UTC m=+92.961152905" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.416957 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=17.416939525 podStartE2EDuration="17.416939525s" podCreationTimestamp="2025-12-10 14:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.415975229 +0000 UTC m=+92.985192869" watchObservedRunningTime="2025-12-10 14:25:43.416939525 +0000 UTC m=+92.986157155" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.449974 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.449955827 podStartE2EDuration="1m15.449955827s" podCreationTimestamp="2025-12-10 14:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:43.434875581 +0000 UTC m=+93.004093221" watchObservedRunningTime="2025-12-10 14:25:43.449955827 +0000 UTC m=+93.019173447" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.488105 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" Dec 10 14:25:43 crc kubenswrapper[4847]: I1210 14:25:43.759491 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:43 crc kubenswrapper[4847]: E1210 14:25:43.759791 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.198914 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" event={"ID":"8e123a0f-959e-42ab-af61-fb62222ae80c","Type":"ContainerStarted","Data":"c1a6f9c8bfd8a27ba1bf4709d5712a624e846b5831c1a9aece5c58d87c8bd589"} Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.198989 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" event={"ID":"8e123a0f-959e-42ab-af61-fb62222ae80c","Type":"ContainerStarted","Data":"dd8f622967ce47923322c1f4c4b9d5d50b37022cc947ec048b401dcc7942972d"} Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.213208 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wk2cf" podStartSLOduration=73.213193582 podStartE2EDuration="1m13.213193582s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:25:44.21275985 +0000 UTC m=+93.781977500" watchObservedRunningTime="2025-12-10 14:25:44.213193582 +0000 UTC m=+93.782411212" Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.759291 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.759382 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:44 crc kubenswrapper[4847]: I1210 14:25:44.759414 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:44 crc kubenswrapper[4847]: E1210 14:25:44.759464 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:44 crc kubenswrapper[4847]: E1210 14:25:44.759536 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:44 crc kubenswrapper[4847]: E1210 14:25:44.759757 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:45 crc kubenswrapper[4847]: I1210 14:25:45.758968 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:45 crc kubenswrapper[4847]: E1210 14:25:45.759196 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:46 crc kubenswrapper[4847]: I1210 14:25:46.758835 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:46 crc kubenswrapper[4847]: I1210 14:25:46.758865 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:46 crc kubenswrapper[4847]: I1210 14:25:46.758949 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:46 crc kubenswrapper[4847]: E1210 14:25:46.759008 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:46 crc kubenswrapper[4847]: E1210 14:25:46.759139 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:46 crc kubenswrapper[4847]: E1210 14:25:46.759206 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:47 crc kubenswrapper[4847]: I1210 14:25:47.758814 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:47 crc kubenswrapper[4847]: E1210 14:25:47.758942 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:48 crc kubenswrapper[4847]: I1210 14:25:48.758995 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:48 crc kubenswrapper[4847]: I1210 14:25:48.759038 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:48 crc kubenswrapper[4847]: E1210 14:25:48.759162 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:48 crc kubenswrapper[4847]: I1210 14:25:48.759218 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:48 crc kubenswrapper[4847]: E1210 14:25:48.759346 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:48 crc kubenswrapper[4847]: E1210 14:25:48.759453 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:49 crc kubenswrapper[4847]: I1210 14:25:49.536555 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:49 crc kubenswrapper[4847]: E1210 14:25:49.536803 4847 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:49 crc kubenswrapper[4847]: E1210 14:25:49.536918 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs podName:e4db231f-2f4a-40e0-8137-c45b3f604fb6 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:53.536886548 +0000 UTC m=+163.106104208 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs") pod "network-metrics-daemon-2khbq" (UID: "e4db231f-2f4a-40e0-8137-c45b3f604fb6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 14:25:49 crc kubenswrapper[4847]: I1210 14:25:49.759310 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:49 crc kubenswrapper[4847]: E1210 14:25:49.759456 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:50 crc kubenswrapper[4847]: I1210 14:25:50.759427 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:50 crc kubenswrapper[4847]: I1210 14:25:50.759564 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:50 crc kubenswrapper[4847]: I1210 14:25:50.760487 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:50 crc kubenswrapper[4847]: E1210 14:25:50.760472 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:50 crc kubenswrapper[4847]: E1210 14:25:50.760664 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:50 crc kubenswrapper[4847]: E1210 14:25:50.760783 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:51 crc kubenswrapper[4847]: I1210 14:25:51.758854 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:51 crc kubenswrapper[4847]: E1210 14:25:51.759077 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:52 crc kubenswrapper[4847]: I1210 14:25:52.759053 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:52 crc kubenswrapper[4847]: E1210 14:25:52.759207 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:52 crc kubenswrapper[4847]: I1210 14:25:52.759253 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:52 crc kubenswrapper[4847]: I1210 14:25:52.759276 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:52 crc kubenswrapper[4847]: E1210 14:25:52.759540 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:52 crc kubenswrapper[4847]: E1210 14:25:52.759602 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:53 crc kubenswrapper[4847]: I1210 14:25:53.759079 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:53 crc kubenswrapper[4847]: E1210 14:25:53.759221 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:53 crc kubenswrapper[4847]: I1210 14:25:53.760744 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:25:53 crc kubenswrapper[4847]: E1210 14:25:53.761074 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-b5l5q_openshift-ovn-kubernetes(9b405b3d-ae75-409a-9e83-b098f333a5c0)\"" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" Dec 10 14:25:54 crc kubenswrapper[4847]: I1210 14:25:54.759358 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:54 crc kubenswrapper[4847]: I1210 14:25:54.759429 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:54 crc kubenswrapper[4847]: I1210 14:25:54.759386 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:54 crc kubenswrapper[4847]: E1210 14:25:54.759516 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:54 crc kubenswrapper[4847]: E1210 14:25:54.759646 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:54 crc kubenswrapper[4847]: E1210 14:25:54.759772 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:55 crc kubenswrapper[4847]: I1210 14:25:55.759300 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:55 crc kubenswrapper[4847]: E1210 14:25:55.759460 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:56 crc kubenswrapper[4847]: I1210 14:25:56.758858 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:56 crc kubenswrapper[4847]: I1210 14:25:56.759004 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:56 crc kubenswrapper[4847]: I1210 14:25:56.759068 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:56 crc kubenswrapper[4847]: E1210 14:25:56.759020 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:56 crc kubenswrapper[4847]: E1210 14:25:56.759240 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:56 crc kubenswrapper[4847]: E1210 14:25:56.759454 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:57 crc kubenswrapper[4847]: I1210 14:25:57.759078 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:57 crc kubenswrapper[4847]: E1210 14:25:57.759218 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:25:58 crc kubenswrapper[4847]: I1210 14:25:58.759547 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:25:58 crc kubenswrapper[4847]: I1210 14:25:58.759607 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:25:58 crc kubenswrapper[4847]: E1210 14:25:58.759754 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:25:58 crc kubenswrapper[4847]: I1210 14:25:58.759578 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:25:58 crc kubenswrapper[4847]: E1210 14:25:58.759910 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:25:58 crc kubenswrapper[4847]: E1210 14:25:58.760075 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:25:59 crc kubenswrapper[4847]: I1210 14:25:59.759316 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:25:59 crc kubenswrapper[4847]: E1210 14:25:59.759471 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:00 crc kubenswrapper[4847]: I1210 14:26:00.759620 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:00 crc kubenswrapper[4847]: I1210 14:26:00.759670 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:00 crc kubenswrapper[4847]: E1210 14:26:00.761484 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:00 crc kubenswrapper[4847]: I1210 14:26:00.761523 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:00 crc kubenswrapper[4847]: E1210 14:26:00.762025 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:00 crc kubenswrapper[4847]: E1210 14:26:00.762218 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:01 crc kubenswrapper[4847]: I1210 14:26:01.759081 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:01 crc kubenswrapper[4847]: E1210 14:26:01.759191 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:02 crc kubenswrapper[4847]: I1210 14:26:02.759582 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:02 crc kubenswrapper[4847]: I1210 14:26:02.759617 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:02 crc kubenswrapper[4847]: E1210 14:26:02.759934 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:02 crc kubenswrapper[4847]: I1210 14:26:02.759981 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:02 crc kubenswrapper[4847]: E1210 14:26:02.760135 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:02 crc kubenswrapper[4847]: E1210 14:26:02.760829 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:03 crc kubenswrapper[4847]: I1210 14:26:03.759042 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:03 crc kubenswrapper[4847]: E1210 14:26:03.759164 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:04 crc kubenswrapper[4847]: I1210 14:26:04.759519 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:04 crc kubenswrapper[4847]: I1210 14:26:04.759523 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:04 crc kubenswrapper[4847]: I1210 14:26:04.759789 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:04 crc kubenswrapper[4847]: E1210 14:26:04.759857 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:04 crc kubenswrapper[4847]: E1210 14:26:04.759690 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:04 crc kubenswrapper[4847]: E1210 14:26:04.760072 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.268953 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/1.log" Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.269746 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/0.log" Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.269854 4847 generic.go:334] "Generic (PLEG): container finished" podID="71681159-7da8-4bc9-837c-d0e3b7397e2e" containerID="c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67" exitCode=1 Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.269964 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerDied","Data":"c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67"} Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.270034 4847 scope.go:117] "RemoveContainer" containerID="b10fd3de8a907fb3da5d12f465b2ccbcc56a02eb1245caff9ab6ec5786cd9a15" Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.270699 4847 scope.go:117] "RemoveContainer" containerID="c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67" Dec 10 14:26:05 crc kubenswrapper[4847]: E1210 14:26:05.270973 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-67b5l_openshift-multus(71681159-7da8-4bc9-837c-d0e3b7397e2e)\"" pod="openshift-multus/multus-67b5l" podUID="71681159-7da8-4bc9-837c-d0e3b7397e2e" Dec 10 14:26:05 crc kubenswrapper[4847]: I1210 14:26:05.759514 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:05 crc kubenswrapper[4847]: E1210 14:26:05.760034 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:06 crc kubenswrapper[4847]: I1210 14:26:06.276530 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/1.log" Dec 10 14:26:06 crc kubenswrapper[4847]: I1210 14:26:06.759951 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:06 crc kubenswrapper[4847]: I1210 14:26:06.759977 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:06 crc kubenswrapper[4847]: I1210 14:26:06.760022 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:06 crc kubenswrapper[4847]: E1210 14:26:06.760999 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:06 crc kubenswrapper[4847]: E1210 14:26:06.761106 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:06 crc kubenswrapper[4847]: E1210 14:26:06.761154 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:07 crc kubenswrapper[4847]: I1210 14:26:07.758984 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:07 crc kubenswrapper[4847]: E1210 14:26:07.759137 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:08 crc kubenswrapper[4847]: I1210 14:26:08.758967 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:08 crc kubenswrapper[4847]: I1210 14:26:08.759053 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:08 crc kubenswrapper[4847]: I1210 14:26:08.759264 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:08 crc kubenswrapper[4847]: E1210 14:26:08.759463 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:08 crc kubenswrapper[4847]: E1210 14:26:08.759844 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:08 crc kubenswrapper[4847]: E1210 14:26:08.759925 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:08 crc kubenswrapper[4847]: I1210 14:26:08.760274 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.286243 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/3.log" Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.288414 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerStarted","Data":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.288923 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.632330 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podStartSLOduration=98.632309282 podStartE2EDuration="1m38.632309282s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:09.34435134 +0000 UTC m=+118.913568990" watchObservedRunningTime="2025-12-10 14:26:09.632309282 +0000 UTC m=+119.201526912" Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.633496 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2khbq"] Dec 10 14:26:09 crc kubenswrapper[4847]: I1210 14:26:09.633614 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:09 crc kubenswrapper[4847]: E1210 14:26:09.633747 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:10 crc kubenswrapper[4847]: E1210 14:26:10.713697 4847 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 10 14:26:10 crc kubenswrapper[4847]: I1210 14:26:10.758618 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:10 crc kubenswrapper[4847]: I1210 14:26:10.758623 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:10 crc kubenswrapper[4847]: I1210 14:26:10.758739 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:10 crc kubenswrapper[4847]: E1210 14:26:10.759698 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:10 crc kubenswrapper[4847]: E1210 14:26:10.759857 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:10 crc kubenswrapper[4847]: E1210 14:26:10.759945 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:10 crc kubenswrapper[4847]: E1210 14:26:10.884604 4847 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 14:26:11 crc kubenswrapper[4847]: I1210 14:26:11.758840 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:11 crc kubenswrapper[4847]: E1210 14:26:11.759058 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:12 crc kubenswrapper[4847]: I1210 14:26:12.759345 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:12 crc kubenswrapper[4847]: I1210 14:26:12.759401 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:12 crc kubenswrapper[4847]: E1210 14:26:12.759495 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:12 crc kubenswrapper[4847]: I1210 14:26:12.759571 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:12 crc kubenswrapper[4847]: E1210 14:26:12.759699 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:12 crc kubenswrapper[4847]: E1210 14:26:12.759807 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:13 crc kubenswrapper[4847]: I1210 14:26:13.759151 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:13 crc kubenswrapper[4847]: E1210 14:26:13.759423 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:14 crc kubenswrapper[4847]: I1210 14:26:14.759502 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:14 crc kubenswrapper[4847]: E1210 14:26:14.759746 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:14 crc kubenswrapper[4847]: I1210 14:26:14.759520 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:14 crc kubenswrapper[4847]: E1210 14:26:14.759862 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:14 crc kubenswrapper[4847]: I1210 14:26:14.759521 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:14 crc kubenswrapper[4847]: E1210 14:26:14.759954 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:15 crc kubenswrapper[4847]: I1210 14:26:15.758870 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:15 crc kubenswrapper[4847]: E1210 14:26:15.759004 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:15 crc kubenswrapper[4847]: E1210 14:26:15.886194 4847 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 14:26:16 crc kubenswrapper[4847]: I1210 14:26:16.758841 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:16 crc kubenswrapper[4847]: I1210 14:26:16.758844 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:16 crc kubenswrapper[4847]: I1210 14:26:16.758968 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:16 crc kubenswrapper[4847]: E1210 14:26:16.759263 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:16 crc kubenswrapper[4847]: I1210 14:26:16.759346 4847 scope.go:117] "RemoveContainer" containerID="c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67" Dec 10 14:26:16 crc kubenswrapper[4847]: E1210 14:26:16.759406 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:16 crc kubenswrapper[4847]: E1210 14:26:16.759467 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:17 crc kubenswrapper[4847]: I1210 14:26:17.759377 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:17 crc kubenswrapper[4847]: E1210 14:26:17.759941 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:18 crc kubenswrapper[4847]: I1210 14:26:18.317146 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/1.log" Dec 10 14:26:18 crc kubenswrapper[4847]: I1210 14:26:18.317200 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerStarted","Data":"a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a"} Dec 10 14:26:18 crc kubenswrapper[4847]: I1210 14:26:18.759117 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:18 crc kubenswrapper[4847]: I1210 14:26:18.759242 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:18 crc kubenswrapper[4847]: I1210 14:26:18.759229 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:18 crc kubenswrapper[4847]: E1210 14:26:18.759406 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:18 crc kubenswrapper[4847]: E1210 14:26:18.759543 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:18 crc kubenswrapper[4847]: E1210 14:26:18.759635 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:19 crc kubenswrapper[4847]: I1210 14:26:19.759621 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:19 crc kubenswrapper[4847]: E1210 14:26:19.759836 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2khbq" podUID="e4db231f-2f4a-40e0-8137-c45b3f604fb6" Dec 10 14:26:20 crc kubenswrapper[4847]: I1210 14:26:20.758809 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:20 crc kubenswrapper[4847]: I1210 14:26:20.758870 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:20 crc kubenswrapper[4847]: I1210 14:26:20.758963 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:20 crc kubenswrapper[4847]: E1210 14:26:20.760054 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 14:26:20 crc kubenswrapper[4847]: E1210 14:26:20.760159 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 14:26:20 crc kubenswrapper[4847]: E1210 14:26:20.760245 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 14:26:21 crc kubenswrapper[4847]: I1210 14:26:21.759008 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:21 crc kubenswrapper[4847]: I1210 14:26:21.761828 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 14:26:21 crc kubenswrapper[4847]: I1210 14:26:21.761880 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.758939 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.758972 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.759065 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.761527 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.761850 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.762049 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 14:26:22 crc kubenswrapper[4847]: I1210 14:26:22.762363 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.390353 4847 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.442399 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jvtvd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.443091 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.445985 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.446848 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.447002 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.447020 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.446858 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.447369 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.447899 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.448021 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.448077 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.448177 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.464959 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.465944 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.466140 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.467178 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-79pvc"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.467466 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.468286 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.468506 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.469053 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.469364 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.473101 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.473551 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.473816 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cmqtl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.474183 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.474433 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6lxgl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.475100 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.475454 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht7zw"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.476050 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.476611 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.478673 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.479207 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.480163 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.480695 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.488937 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.512831 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.525479 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.525659 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526215 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526316 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526353 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526464 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526575 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526726 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526942 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527127 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527239 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527325 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527371 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527410 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-trusted-ca\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527448 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df9cc62d-5823-4006-b5ec-ce6c9c31f560-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527486 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjj5h\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-kube-api-access-jjj5h\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527510 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxnr6\" (UniqueName: \"kubernetes.io/projected/57fb36d2-42fe-4993-a454-1db53581dc35-kube-api-access-gxnr6\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527584 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527634 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit-dir\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527663 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8bxk\" (UniqueName: \"kubernetes.io/projected/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-kube-api-access-w8bxk\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527704 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g8bg\" (UniqueName: \"kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527737 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtvkn\" (UniqueName: \"kubernetes.io/projected/074675ba-69db-4df1-b8dd-d05912a3588a-kube-api-access-mtvkn\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527766 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f3e2aa2-3012-45f6-b564-97e4be748467-machine-approver-tls\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527791 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527812 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv5z5\" (UniqueName: \"kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527842 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527872 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527888 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527905 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527931 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527966 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-client\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527988 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-config\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528012 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-serving-cert\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528032 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528056 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-auth-proxy-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528075 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-config\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528096 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528123 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-config\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528146 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528168 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqgdr\" (UniqueName: \"kubernetes.io/projected/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-kube-api-access-zqgdr\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528186 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528205 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528221 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b3fc300-4e72-4e72-9ff2-ad647815fac8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528238 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528258 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528282 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528312 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528326 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528343 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528358 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528376 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528393 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528431 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-service-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528451 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528478 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-image-import-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528514 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528549 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-images\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528575 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528603 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df9cc62d-5823-4006-b5ec-ce6c9c31f560-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528641 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528681 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b3fc300-4e72-4e72-9ff2-ad647815fac8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528701 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528739 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528774 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def56373-569e-4262-8cbb-9a2d880de58d-serving-cert\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528790 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjftv\" (UniqueName: \"kubernetes.io/projected/def56373-569e-4262-8cbb-9a2d880de58d-kube-api-access-jjftv\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528808 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmnk9\" (UniqueName: \"kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528821 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/57fb36d2-42fe-4993-a454-1db53581dc35-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528860 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5jsc\" (UniqueName: \"kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528877 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528908 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp8vk\" (UniqueName: \"kubernetes.io/projected/4b3fc300-4e72-4e72-9ff2-ad647815fac8-kube-api-access-jp8vk\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528924 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-encryption-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528939 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528953 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc8t2\" (UniqueName: \"kubernetes.io/projected/1f3e2aa2-3012-45f6-b564-97e4be748467-kube-api-access-gc8t2\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528968 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529015 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-node-pullsecrets\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529032 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529047 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074675ba-69db-4df1-b8dd-d05912a3588a-serving-cert\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529062 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529084 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529098 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.525524 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527424 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532435 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532468 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.533555 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.533810 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526318 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527149 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527566 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526266 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527566 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527750 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.534640 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527755 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527782 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527786 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527826 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527870 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527903 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.527947 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528057 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528072 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528115 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528201 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528373 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528421 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528458 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528502 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.528559 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.529390 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530495 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530564 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530762 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530798 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530816 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530866 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530868 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530914 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.530972 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.526247 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531012 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531028 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531057 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531080 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531125 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531162 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531063 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531228 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.531805 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532047 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532071 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532137 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532173 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.532233 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.536928 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.540016 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.547192 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.539574 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-c6lv4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.550918 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.551076 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.551218 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.551321 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.551423 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.557037 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.559693 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.561218 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.562973 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.551231 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmc4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.563894 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.564061 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.581604 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.584264 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.585130 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-csskz"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.586125 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.587523 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wj5h"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.587778 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.606876 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.607621 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.608211 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.608357 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612186 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612360 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612482 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612550 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612564 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612565 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612632 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.612925 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.613092 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.613145 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.613193 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.613321 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.618311 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pg8lp"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.618404 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.618805 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.618916 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.620795 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.621231 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.625638 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.626058 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.626178 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.626388 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.626621 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630297 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630331 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630360 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgs47\" (UniqueName: \"kubernetes.io/projected/10f6f206-7866-45bf-a629-0eb1f832cda4-kube-api-access-fgs47\") pod \"downloads-7954f5f757-c6lv4\" (UID: \"10f6f206-7866-45bf-a629-0eb1f832cda4\") " pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630376 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630394 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/176f2b83-e66f-43af-8354-69f2ee03315a-metrics-tls\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630412 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630427 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x68qs\" (UniqueName: \"kubernetes.io/projected/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-kube-api-access-x68qs\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630443 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630458 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-trusted-ca\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630473 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjj5h\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-kube-api-access-jjj5h\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630490 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxnr6\" (UniqueName: \"kubernetes.io/projected/57fb36d2-42fe-4993-a454-1db53581dc35-kube-api-access-gxnr6\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630507 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df9cc62d-5823-4006-b5ec-ce6c9c31f560-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630523 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8bxk\" (UniqueName: \"kubernetes.io/projected/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-kube-api-access-w8bxk\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630539 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g8bg\" (UniqueName: \"kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630555 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630573 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit-dir\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630587 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtvkn\" (UniqueName: \"kubernetes.io/projected/074675ba-69db-4df1-b8dd-d05912a3588a-kube-api-access-mtvkn\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630603 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8123d07a-88ab-42d4-b554-1a6062fec387-proxy-tls\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630629 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f3e2aa2-3012-45f6-b564-97e4be748467-machine-approver-tls\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630673 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630762 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630783 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630807 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv5z5\" (UniqueName: \"kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630843 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630858 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630875 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-client\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630890 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630914 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-config\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630932 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b826590-f030-4a21-a7b2-ac77deda13e5-serving-cert\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630947 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-serving-cert\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630965 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630982 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-auth-proxy-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.630997 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-config\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631012 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-config\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631028 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631045 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9b826590-f030-4a21-a7b2-ac77deda13e5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631064 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631084 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqgdr\" (UniqueName: \"kubernetes.io/projected/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-kube-api-access-zqgdr\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631101 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631116 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631133 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b3fc300-4e72-4e72-9ff2-ad647815fac8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631148 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631166 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631182 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-metrics-tls\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631198 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwvkh\" (UniqueName: \"kubernetes.io/projected/9b826590-f030-4a21-a7b2-ac77deda13e5-kube-api-access-cwvkh\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631213 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrhp\" (UniqueName: \"kubernetes.io/projected/8123d07a-88ab-42d4-b554-1a6062fec387-kube-api-access-hgrhp\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631230 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631252 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631267 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631283 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-service-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631300 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631315 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631403 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631418 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631434 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szmps\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-kube-api-access-szmps\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631450 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631464 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-image-import-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631478 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631493 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-images\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631511 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/176f2b83-e66f-43af-8354-69f2ee03315a-trusted-ca\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631527 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df9cc62d-5823-4006-b5ec-ce6c9c31f560-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631541 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631556 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631573 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b3fc300-4e72-4e72-9ff2-ad647815fac8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631588 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbxtd\" (UniqueName: \"kubernetes.io/projected/9fc9d601-18de-40ac-b15d-6d385f61124e-kube-api-access-jbxtd\") pod \"migrator-59844c95c7-t95rl\" (UID: \"9fc9d601-18de-40ac-b15d-6d385f61124e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631603 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631648 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631664 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def56373-569e-4262-8cbb-9a2d880de58d-serving-cert\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631736 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjftv\" (UniqueName: \"kubernetes.io/projected/def56373-569e-4262-8cbb-9a2d880de58d-kube-api-access-jjftv\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631754 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmnk9\" (UniqueName: \"kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631769 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/57fb36d2-42fe-4993-a454-1db53581dc35-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631792 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5jsc\" (UniqueName: \"kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631809 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8123d07a-88ab-42d4-b554-1a6062fec387-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631824 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631840 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp8vk\" (UniqueName: \"kubernetes.io/projected/4b3fc300-4e72-4e72-9ff2-ad647815fac8-kube-api-access-jp8vk\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631856 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-encryption-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631895 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc8t2\" (UniqueName: \"kubernetes.io/projected/1f3e2aa2-3012-45f6-b564-97e4be748467-kube-api-access-gc8t2\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631912 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631933 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-node-pullsecrets\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631932 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-auth-proxy-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.631949 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.632014 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074675ba-69db-4df1-b8dd-d05912a3588a-serving-cert\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.632526 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.632748 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.633215 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-image-import-ca\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.633524 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-config\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.633970 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.634292 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-78j8d"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.634309 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-config\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.634521 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/57fb36d2-42fe-4993-a454-1db53581dc35-images\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.634705 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.634844 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-config\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.635011 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.635542 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.635940 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.637452 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-audit-dir\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.636127 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.639855 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.640351 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.640556 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.640623 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.640925 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-serving-cert\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.640959 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.641190 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-etcd-client\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.641281 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.641328 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.641338 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.641679 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-node-pullsecrets\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642029 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642142 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642190 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642594 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642620 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/def56373-569e-4262-8cbb-9a2d880de58d-service-ca-bundle\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.642836 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.643943 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.644053 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.644185 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.644189 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.644300 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.644322 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.647406 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.647663 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.647726 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.647800 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.647884 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.648113 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/57fb36d2-42fe-4993-a454-1db53581dc35-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.648166 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1f3e2aa2-3012-45f6-b564-97e4be748467-machine-approver-tls\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.648290 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.648311 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.648397 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649080 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649305 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649422 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649554 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f3e2aa2-3012-45f6-b564-97e4be748467-config\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649733 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649913 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b3fc300-4e72-4e72-9ff2-ad647815fac8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650187 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/074675ba-69db-4df1-b8dd-d05912a3588a-trusted-ca\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650292 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.649548 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650498 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650602 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650600 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650909 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b3fc300-4e72-4e72-9ff2-ad647815fac8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.650922 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df9cc62d-5823-4006-b5ec-ce6c9c31f560-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651524 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651676 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651866 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651899 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651774 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651957 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.652018 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/df9cc62d-5823-4006-b5ec-ce6c9c31f560-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.651677 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-encryption-config\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.652538 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.652631 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.653052 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/def56373-569e-4262-8cbb-9a2d880de58d-serving-cert\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.653254 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.653814 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g485q"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.654245 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.654852 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/074675ba-69db-4df1-b8dd-d05912a3588a-serving-cert\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.654978 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.655062 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.660491 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.661356 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.661934 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.662269 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.663407 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.663452 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cmqtl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.663469 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.665446 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-79pvc"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.664525 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.668813 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbdjm"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.670502 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.670528 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6lxgl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.670667 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.671986 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-csskz"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.672554 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.674783 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.676642 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wj5h"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.677936 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.678921 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.680178 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht7zw"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.681167 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.683297 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmc4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.684395 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jvtvd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.684830 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.685374 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.686529 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.687893 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.688406 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c6lv4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.689351 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.690312 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pg8lp"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.691260 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.692198 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.693099 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.694024 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-mzz74"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.694628 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.694930 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6qqn9"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.695740 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.695884 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.696821 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.697753 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.698749 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.699750 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.700721 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.701751 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.703979 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.705352 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.705379 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6qqn9"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.706679 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.708107 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mzz74"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.709266 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbdjm"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.710343 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g485q"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.711380 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.712450 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-h6hmg"] Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.713533 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.726704 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733302 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-mountpoint-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733341 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733362 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733391 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgs47\" (UniqueName: \"kubernetes.io/projected/10f6f206-7866-45bf-a629-0eb1f832cda4-kube-api-access-fgs47\") pod \"downloads-7954f5f757-c6lv4\" (UID: \"10f6f206-7866-45bf-a629-0eb1f832cda4\") " pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733412 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/176f2b83-e66f-43af-8354-69f2ee03315a-metrics-tls\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733431 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x68qs\" (UniqueName: \"kubernetes.io/projected/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-kube-api-access-x68qs\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733448 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733504 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8123d07a-88ab-42d4-b554-1a6062fec387-proxy-tls\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733519 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-plugins-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733542 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-dir\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733581 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b826590-f030-4a21-a7b2-ac77deda13e5-serving-cert\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733598 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733613 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjkpj\" (UniqueName: \"kubernetes.io/projected/420a02dc-3c91-4c2c-b885-06fa71cb2e17-kube-api-access-cjkpj\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733631 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9b826590-f030-4a21-a7b2-ac77deda13e5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733654 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733675 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chr98\" (UniqueName: \"kubernetes.io/projected/be8157de-bba2-44f2-b301-031caf2f46f5-kube-api-access-chr98\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733692 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xsjz\" (UniqueName: \"kubernetes.io/projected/c1857b2d-62af-42c8-94c7-6fa4c248e521-kube-api-access-5xsjz\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733724 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-metrics-tls\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733741 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwvkh\" (UniqueName: \"kubernetes.io/projected/9b826590-f030-4a21-a7b2-ac77deda13e5-kube-api-access-cwvkh\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733760 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrhp\" (UniqueName: \"kubernetes.io/projected/8123d07a-88ab-42d4-b554-1a6062fec387-kube-api-access-hgrhp\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733781 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733796 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733819 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/176f2b83-e66f-43af-8354-69f2ee03315a-trusted-ca\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733838 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szmps\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-kube-api-access-szmps\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733859 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733875 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-socket-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733907 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbxtd\" (UniqueName: \"kubernetes.io/projected/9fc9d601-18de-40ac-b15d-6d385f61124e-kube-api-access-jbxtd\") pod \"migrator-59844c95c7-t95rl\" (UID: \"9fc9d601-18de-40ac-b15d-6d385f61124e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733927 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-registration-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733942 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733970 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-csi-data-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.733992 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8123d07a-88ab-42d4-b554-1a6062fec387-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.734208 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9b826590-f030-4a21-a7b2-ac77deda13e5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.735067 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8123d07a-88ab-42d4-b554-1a6062fec387-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.735224 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/176f2b83-e66f-43af-8354-69f2ee03315a-trusted-ca\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.737052 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-metrics-tls\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.737201 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8123d07a-88ab-42d4-b554-1a6062fec387-proxy-tls\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.737886 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b826590-f030-4a21-a7b2-ac77deda13e5-serving-cert\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.738734 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/176f2b83-e66f-43af-8354-69f2ee03315a-metrics-tls\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.746805 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.768244 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.785311 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.805181 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.832141 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834556 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834624 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-plugins-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834649 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-dir\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834688 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834706 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjkpj\" (UniqueName: \"kubernetes.io/projected/420a02dc-3c91-4c2c-b885-06fa71cb2e17-kube-api-access-cjkpj\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834747 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834766 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chr98\" (UniqueName: \"kubernetes.io/projected/be8157de-bba2-44f2-b301-031caf2f46f5-kube-api-access-chr98\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834785 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xsjz\" (UniqueName: \"kubernetes.io/projected/c1857b2d-62af-42c8-94c7-6fa4c248e521-kube-api-access-5xsjz\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834813 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834830 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834870 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834886 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-socket-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834905 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834923 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-registration-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834946 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-csi-data-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.834991 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-mountpoint-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835013 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835008 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-plugins-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835014 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-dir\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835101 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-registration-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835162 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-csi-data-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835168 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-mountpoint-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.835162 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/be8157de-bba2-44f2-b301-031caf2f46f5-socket-dir\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.844947 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.865538 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.885073 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.905295 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.925595 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.945075 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.965610 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 14:26:24 crc kubenswrapper[4847]: I1210 14:26:24.985531 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.005089 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.025009 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.045976 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.066467 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.086009 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.105505 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.145732 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv5z5\" (UniqueName: \"kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5\") pod \"route-controller-manager-6576b87f9c-b7rjd\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.160939 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.179079 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjj5h\" (UniqueName: \"kubernetes.io/projected/df9cc62d-5823-4006-b5ec-ce6c9c31f560-kube-api-access-jjj5h\") pod \"cluster-image-registry-operator-dc59b4c8b-tmhm6\" (UID: \"df9cc62d-5823-4006-b5ec-ce6c9c31f560\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.185053 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.205272 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.225475 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.230201 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.246102 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.265560 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.300479 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqgdr\" (UniqueName: \"kubernetes.io/projected/b8a2531c-797a-4b45-8e8b-aa2b278c7d49-kube-api-access-zqgdr\") pod \"openshift-controller-manager-operator-756b6f6bc6-ml27d\" (UID: \"b8a2531c-797a-4b45-8e8b-aa2b278c7d49\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.320483 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxnr6\" (UniqueName: \"kubernetes.io/projected/57fb36d2-42fe-4993-a454-1db53581dc35-kube-api-access-gxnr6\") pod \"machine-api-operator-5694c8668f-6lxgl\" (UID: \"57fb36d2-42fe-4993-a454-1db53581dc35\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.341947 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5jsc\" (UniqueName: \"kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc\") pod \"controller-manager-879f6c89f-5xqhd\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.359464 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8bxk\" (UniqueName: \"kubernetes.io/projected/6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e-kube-api-access-w8bxk\") pod \"apiserver-76f77b778f-jvtvd\" (UID: \"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e\") " pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.369027 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.379982 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g8bg\" (UniqueName: \"kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg\") pod \"console-f9d7485db-m2sr6\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.390116 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6"] Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.399685 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtvkn\" (UniqueName: \"kubernetes.io/projected/074675ba-69db-4df1-b8dd-d05912a3588a-kube-api-access-mtvkn\") pod \"console-operator-58897d9998-cmqtl\" (UID: \"074675ba-69db-4df1-b8dd-d05912a3588a\") " pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.401653 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf9cc62d_5823_4006_b5ec_ce6c9c31f560.slice/crio-db744c6252e8501cc33c69f6e8f3bc6efb6280ed4166269a337f3cb2f8ad3a53 WatchSource:0}: Error finding container db744c6252e8501cc33c69f6e8f3bc6efb6280ed4166269a337f3cb2f8ad3a53: Status 404 returned error can't find the container with id db744c6252e8501cc33c69f6e8f3bc6efb6280ed4166269a337f3cb2f8ad3a53 Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.421767 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp8vk\" (UniqueName: \"kubernetes.io/projected/4b3fc300-4e72-4e72-9ff2-ad647815fac8-kube-api-access-jp8vk\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cbh\" (UID: \"4b3fc300-4e72-4e72-9ff2-ad647815fac8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.426203 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.432135 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.460726 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc8t2\" (UniqueName: \"kubernetes.io/projected/1f3e2aa2-3012-45f6-b564-97e4be748467-kube-api-access-gc8t2\") pod \"machine-approver-56656f9798-74bhj\" (UID: \"1f3e2aa2-3012-45f6-b564-97e4be748467\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.465876 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.485830 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.488679 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.506987 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.512920 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f3e2aa2_3012_45f6_b564_97e4be748467.slice/crio-997f3972074670831894d761e6221955a30638bfb0093ced6af3aa20a3db34bb WatchSource:0}: Error finding container 997f3972074670831894d761e6221955a30638bfb0093ced6af3aa20a3db34bb: Status 404 returned error can't find the container with id 997f3972074670831894d761e6221955a30638bfb0093ced6af3aa20a3db34bb Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.518307 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.524879 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.534081 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jvtvd"] Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.535666 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.545411 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b873c5d_e3ab_41b1_a8ba_10aa5b446c9e.slice/crio-987e91461810fc56c7b72f2b45a467d5b89d6a3828230a99e9b8d53ee9b35f00 WatchSource:0}: Error finding container 987e91461810fc56c7b72f2b45a467d5b89d6a3828230a99e9b8d53ee9b35f00: Status 404 returned error can't find the container with id 987e91461810fc56c7b72f2b45a467d5b89d6a3828230a99e9b8d53ee9b35f00 Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.565799 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.566402 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjftv\" (UniqueName: \"kubernetes.io/projected/def56373-569e-4262-8cbb-9a2d880de58d-kube-api-access-jjftv\") pod \"authentication-operator-69f744f599-79pvc\" (UID: \"def56373-569e-4262-8cbb-9a2d880de58d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.586061 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.597647 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.604178 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.618159 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.623292 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmnk9\" (UniqueName: \"kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9\") pod \"oauth-openshift-558db77b4-ht7zw\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.625856 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.644982 4847 request.go:700] Waited for 1.000556059s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/secrets?fieldSelector=metadata.name%3Dkube-apiserver-operator-serving-cert&limit=500&resourceVersion=0 Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.647013 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.665775 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.686075 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.706490 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.712115 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.724907 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.745116 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.751056 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.765933 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.786162 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.805783 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.824352 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835600 4847 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835617 4847 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835607 4847 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835657 4847 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835708 4847 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835674 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335654709 +0000 UTC m=+135.904872339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835774 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls podName:420a02dc-3c91-4c2c-b885-06fa71cb2e17 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335751712 +0000 UTC m=+135.904969342 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls") pod "dns-default-6qqn9" (UID: "420a02dc-3c91-4c2c-b885-06fa71cb2e17") : failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835775 4847 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835787 4847 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835787 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335780523 +0000 UTC m=+135.904998143 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835825 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335802093 +0000 UTC m=+135.905019723 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835843 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335836014 +0000 UTC m=+135.905053644 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835856 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335850195 +0000 UTC m=+135.905067825 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.835869 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client podName:c1857b2d-62af-42c8-94c7-6fa4c248e521 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.335862765 +0000 UTC m=+135.905080395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client") pod "apiserver-7bbb656c7d-fjjbb" (UID: "c1857b2d-62af-42c8-94c7-6fa4c248e521") : failed to sync secret cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.836288 4847 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: E1210 14:26:25.836478 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume podName:420a02dc-3c91-4c2c-b885-06fa71cb2e17 nodeName:}" failed. No retries permitted until 2025-12-10 14:26:26.336461211 +0000 UTC m=+135.905678951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume") pod "dns-default-6qqn9" (UID: "420a02dc-3c91-4c2c-b885-06fa71cb2e17") : failed to sync configmap cache: timed out waiting for the condition Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.844817 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.870670 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.885008 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.892136 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.895793 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.899228 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cmqtl"] Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.903483 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d"] Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.904927 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.926078 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.930012 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod074675ba_69db_4df1_b8dd_d05912a3588a.slice/crio-fe3445aebfddc1e79dfe454fb63526e2e2e330f17c8f84d4296856a7c0e75ee6 WatchSource:0}: Error finding container fe3445aebfddc1e79dfe454fb63526e2e2e330f17c8f84d4296856a7c0e75ee6: Status 404 returned error can't find the container with id fe3445aebfddc1e79dfe454fb63526e2e2e330f17c8f84d4296856a7c0e75ee6 Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.932267 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8a2531c_797a_4b45_8e8b_aa2b278c7d49.slice/crio-388e278fdea750631632869ff3dfeb8afc465caf75a80cc06e6453549cabb910 WatchSource:0}: Error finding container 388e278fdea750631632869ff3dfeb8afc465caf75a80cc06e6453549cabb910: Status 404 returned error can't find the container with id 388e278fdea750631632869ff3dfeb8afc465caf75a80cc06e6453549cabb910 Dec 10 14:26:25 crc kubenswrapper[4847]: W1210 14:26:25.932507 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod598e635e_4f60_4518_9021_456857be17e6.slice/crio-d164c32a972d75c113ad2f222624920c8d519e4f61a82d616a40f1c82450ef43 WatchSource:0}: Error finding container d164c32a972d75c113ad2f222624920c8d519e4f61a82d616a40f1c82450ef43: Status 404 returned error can't find the container with id d164c32a972d75c113ad2f222624920c8d519e4f61a82d616a40f1c82450ef43 Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.945261 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.966829 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 14:26:25 crc kubenswrapper[4847]: I1210 14:26:25.986826 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.005963 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.036840 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.045493 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.067012 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.084836 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.105553 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.125341 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.145271 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.165794 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.187046 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.205693 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.226359 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.245627 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.271473 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.292236 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.304964 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.326750 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.346039 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.361107 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" event={"ID":"598e635e-4f60-4518-9021-456857be17e6","Type":"ContainerStarted","Data":"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.361174 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" event={"ID":"598e635e-4f60-4518-9021-456857be17e6","Type":"ContainerStarted","Data":"d164c32a972d75c113ad2f222624920c8d519e4f61a82d616a40f1c82450ef43"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.362912 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.369746 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e" containerID="e936132f3d34ddbd3bb8992a27720087e345c69c35ae084da04922282fb0829e" exitCode=0 Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.369822 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" event={"ID":"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e","Type":"ContainerDied","Data":"e936132f3d34ddbd3bb8992a27720087e345c69c35ae084da04922282fb0829e"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.369854 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" event={"ID":"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e","Type":"ContainerStarted","Data":"987e91461810fc56c7b72f2b45a467d5b89d6a3828230a99e9b8d53ee9b35f00"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.370937 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.372390 4847 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-b7rjd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.372446 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" podUID="598e635e-4f60-4518-9021-456857be17e6" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376397 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376451 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376507 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376532 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376576 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376643 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376702 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.376756 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.377936 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.378041 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.379336 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" event={"ID":"df9cc62d-5823-4006-b5ec-ce6c9c31f560","Type":"ContainerStarted","Data":"3fbf4776261c7a941c89e67edc5d9458b5dfe1181f7d28161ccec9a7158fbf18"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.379386 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" event={"ID":"df9cc62d-5823-4006-b5ec-ce6c9c31f560","Type":"ContainerStarted","Data":"db744c6252e8501cc33c69f6e8f3bc6efb6280ed4166269a337f3cb2f8ad3a53"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.382511 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-79pvc"] Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.384343 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" event={"ID":"074675ba-69db-4df1-b8dd-d05912a3588a","Type":"ContainerStarted","Data":"13a5f136dba1a81a0e2337f6d656421ee7ecefd0a15245c2090d158c8283c781"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.384396 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" event={"ID":"074675ba-69db-4df1-b8dd-d05912a3588a","Type":"ContainerStarted","Data":"fe3445aebfddc1e79dfe454fb63526e2e2e330f17c8f84d4296856a7c0e75ee6"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.385105 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.387155 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" event={"ID":"b8a2531c-797a-4b45-8e8b-aa2b278c7d49","Type":"ContainerStarted","Data":"fc60f288f293102aa2f4d613ae23616b9eb76cebff4219f1b2e360020fa1ae2d"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.387182 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" event={"ID":"b8a2531c-797a-4b45-8e8b-aa2b278c7d49","Type":"ContainerStarted","Data":"388e278fdea750631632869ff3dfeb8afc465caf75a80cc06e6453549cabb910"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.388441 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" event={"ID":"1f3e2aa2-3012-45f6-b564-97e4be748467","Type":"ContainerStarted","Data":"1abf3c1f9da81401b16c54568f24eb012ed3e54650bf5ff1eebb7a797954eeb6"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.388466 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" event={"ID":"1f3e2aa2-3012-45f6-b564-97e4be748467","Type":"ContainerStarted","Data":"ea3a393d51b8a2399323785fda5685c29556221554e335be884804668670541f"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.388476 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" event={"ID":"1f3e2aa2-3012-45f6-b564-97e4be748467","Type":"ContainerStarted","Data":"997f3972074670831894d761e6221955a30638bfb0093ced6af3aa20a3db34bb"} Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.390299 4847 patch_prober.go:28] interesting pod/console-operator-58897d9998-cmqtl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.390359 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" podUID="074675ba-69db-4df1-b8dd-d05912a3588a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.391934 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.405591 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.415601 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-etcd-client\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.426636 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6lxgl"] Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.428109 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.441829 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-serving-cert\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.448145 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.463530 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c1857b2d-62af-42c8-94c7-6fa4c248e521-encryption-config\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.467500 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.477431 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.485510 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.491706 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht7zw"] Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.494442 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh"] Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.506157 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.518561 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:26:26 crc kubenswrapper[4847]: W1210 14:26:26.520927 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b3fc300_4e72_4e72_9ff2_ad647815fac8.slice/crio-017d1cf81fd5e6458a326ec111aef2dd364f98c779acd85ca72489efc03b269a WatchSource:0}: Error finding container 017d1cf81fd5e6458a326ec111aef2dd364f98c779acd85ca72489efc03b269a: Status 404 returned error can't find the container with id 017d1cf81fd5e6458a326ec111aef2dd364f98c779acd85ca72489efc03b269a Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.540749 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.547630 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c1857b2d-62af-42c8-94c7-6fa4c248e521-audit-policies\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.552499 4847 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 14:26:26 crc kubenswrapper[4847]: W1210 14:26:26.554558 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24d0c394_4cf4_452f_a117_215657acb8f2.slice/crio-207d1362d360284dbb7f224a9bfd265caf3c850c58b071ef7b4e5a84d072331b WatchSource:0}: Error finding container 207d1362d360284dbb7f224a9bfd265caf3c850c58b071ef7b4e5a84d072331b: Status 404 returned error can't find the container with id 207d1362d360284dbb7f224a9bfd265caf3c850c58b071ef7b4e5a84d072331b Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.567160 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.585442 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.628895 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.645781 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.669327 4847 request.go:700] Waited for 1.974451275s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.670942 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.686877 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.705296 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.730221 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.750266 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/420a02dc-3c91-4c2c-b885-06fa71cb2e17-metrics-tls\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.752286 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.758668 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/420a02dc-3c91-4c2c-b885-06fa71cb2e17-config-volume\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.791167 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.808232 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.826514 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.867311 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.881007 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgs47\" (UniqueName: \"kubernetes.io/projected/10f6f206-7866-45bf-a629-0eb1f832cda4-kube-api-access-fgs47\") pod \"downloads-7954f5f757-c6lv4\" (UID: \"10f6f206-7866-45bf-a629-0eb1f832cda4\") " pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.910736 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x68qs\" (UniqueName: \"kubernetes.io/projected/a0fed061-99e2-4dc2-98cb-a81937b7d7b8-kube-api-access-x68qs\") pod \"dns-operator-744455d44c-ppmc4\" (UID: \"a0fed061-99e2-4dc2-98cb-a81937b7d7b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.924877 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrhp\" (UniqueName: \"kubernetes.io/projected/8123d07a-88ab-42d4-b554-1a6062fec387-kube-api-access-hgrhp\") pod \"machine-config-controller-84d6567774-s58tq\" (UID: \"8123d07a-88ab-42d4-b554-1a6062fec387\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.945849 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwvkh\" (UniqueName: \"kubernetes.io/projected/9b826590-f030-4a21-a7b2-ac77deda13e5-kube-api-access-cwvkh\") pod \"openshift-config-operator-7777fb866f-csskz\" (UID: \"9b826590-f030-4a21-a7b2-ac77deda13e5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.964904 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szmps\" (UniqueName: \"kubernetes.io/projected/176f2b83-e66f-43af-8354-69f2ee03315a-kube-api-access-szmps\") pod \"ingress-operator-5b745b69d9-f8dtg\" (UID: \"176f2b83-e66f-43af-8354-69f2ee03315a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:26 crc kubenswrapper[4847]: I1210 14:26:26.982881 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbxtd\" (UniqueName: \"kubernetes.io/projected/9fc9d601-18de-40ac-b15d-6d385f61124e-kube-api-access-jbxtd\") pod \"migrator-59844c95c7-t95rl\" (UID: \"9fc9d601-18de-40ac-b15d-6d385f61124e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.003619 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chr98\" (UniqueName: \"kubernetes.io/projected/be8157de-bba2-44f2-b301-031caf2f46f5-kube-api-access-chr98\") pod \"csi-hostpathplugin-fbdjm\" (UID: \"be8157de-bba2-44f2-b301-031caf2f46f5\") " pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.022073 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjkpj\" (UniqueName: \"kubernetes.io/projected/420a02dc-3c91-4c2c-b885-06fa71cb2e17-kube-api-access-cjkpj\") pod \"dns-default-6qqn9\" (UID: \"420a02dc-3c91-4c2c-b885-06fa71cb2e17\") " pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.044331 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xsjz\" (UniqueName: \"kubernetes.io/projected/c1857b2d-62af-42c8-94c7-6fa4c248e521-kube-api-access-5xsjz\") pod \"apiserver-7bbb656c7d-fjjbb\" (UID: \"c1857b2d-62af-42c8-94c7-6fa4c248e521\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.049240 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.069234 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.084180 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.091788 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.091840 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2wl\" (UniqueName: \"kubernetes.io/projected/5bc4e7d4-84ac-4ab8-ba99-143340209459-kube-api-access-5s2wl\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.091922 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-client\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092005 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-metrics-certs\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe2f13e-8873-45a6-a5c8-b03affdad616-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092084 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-srv-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092142 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82bd8\" (UniqueName: \"kubernetes.io/projected/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-kube-api-access-82bd8\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092202 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zf58\" (UniqueName: \"kubernetes.io/projected/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-kube-api-access-7zf58\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092304 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78175b85-3b98-40b5-aa1a-abc5804132e9-serving-cert\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092371 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-srv-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092402 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctnf8\" (UniqueName: \"kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092430 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rgwm\" (UniqueName: \"kubernetes.io/projected/8671ff24-9cdd-4013-9e9e-f885c5029f58-kube-api-access-5rgwm\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092490 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftpqh\" (UniqueName: \"kubernetes.io/projected/5104e975-a44a-4068-902e-589d3fce2b99-kube-api-access-ftpqh\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092529 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e61140b-3686-4f1b-b8b3-56878044f9f1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092559 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-key\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092633 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe2f13e-8873-45a6-a5c8-b03affdad616-config\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092663 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-service-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092691 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6j76\" (UniqueName: \"kubernetes.io/projected/160c8f8d-4bee-4f69-9051-ee22f8014444-kube-api-access-x6j76\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092747 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092772 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4kqn\" (UniqueName: \"kubernetes.io/projected/78175b85-3b98-40b5-aa1a-abc5804132e9-kube-api-access-r4kqn\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092799 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8671ff24-9cdd-4013-9e9e-f885c5029f58-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092860 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.092928 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093003 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093036 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093061 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-config\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093085 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd54a637-b564-42a5-ae67-2daeda3d2f28-tmpfs\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093111 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-apiservice-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093132 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093155 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxp9z\" (UniqueName: \"kubernetes.io/projected/c95c42c9-7795-4e54-8b28-357ad9aa537e-kube-api-access-fxp9z\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093182 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-webhook-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093224 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjj24\" (UniqueName: \"kubernetes.io/projected/89217e6c-9f20-45f0-8a26-7c808efc3359-kube-api-access-fjj24\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093261 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-cabundle\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093314 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093389 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxbbw\" (UniqueName: \"kubernetes.io/projected/7a8b9b0c-61c8-460d-8226-7794483e1cb1-kube-api-access-kxbbw\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093421 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-service-ca-bundle\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093463 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093486 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093511 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093534 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7a8b9b0c-61c8-460d-8226-7794483e1cb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093559 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4rvf\" (UniqueName: \"kubernetes.io/projected/cd54a637-b564-42a5-ae67-2daeda3d2f28-kube-api-access-z4rvf\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093582 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-default-certificate\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093604 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093626 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093660 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klx7g\" (UniqueName: \"kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093684 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-profile-collector-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093735 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c95c42c9-7795-4e54-8b28-357ad9aa537e-cert\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093785 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe2f13e-8873-45a6-a5c8-b03affdad616-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.093818 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094084 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z4mq\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094173 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094196 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-stats-auth\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094263 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5104e975-a44a-4068-902e-589d3fce2b99-proxy-tls\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094282 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-serving-cert\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094314 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f24mk\" (UniqueName: \"kubernetes.io/projected/d21a2da8-a31f-40a3-84b5-db9d331e822e-kube-api-access-f24mk\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094338 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094413 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-images\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094457 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-config\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094477 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e61140b-3686-4f1b-b8b3-56878044f9f1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094521 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094598 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21a2da8-a31f-40a3-84b5-db9d331e822e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094620 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e61140b-3686-4f1b-b8b3-56878044f9f1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094695 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9h8s\" (UniqueName: \"kubernetes.io/projected/ada1191a-da37-466d-a7cc-a7d97b9f78d4-kube-api-access-t9h8s\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094766 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvdjc\" (UniqueName: \"kubernetes.io/projected/307c7632-1a40-49ae-96fc-66cf82b9e8eb-kube-api-access-rvdjc\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094957 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.094984 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78175b85-3b98-40b5-aa1a-abc5804132e9-config\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.095047 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/89217e6c-9f20-45f0-8a26-7c808efc3359-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.096160 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:27.596133382 +0000 UTC m=+137.165351192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.125000 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.131477 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.154770 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.162999 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.171410 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.178521 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196017 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196199 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-key\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196231 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe2f13e-8873-45a6-a5c8-b03affdad616-config\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196257 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-service-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196281 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6j76\" (UniqueName: \"kubernetes.io/projected/160c8f8d-4bee-4f69-9051-ee22f8014444-kube-api-access-x6j76\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196303 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196325 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4kqn\" (UniqueName: \"kubernetes.io/projected/78175b85-3b98-40b5-aa1a-abc5804132e9-kube-api-access-r4kqn\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196350 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8671ff24-9cdd-4013-9e9e-f885c5029f58-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196376 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196398 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196421 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196443 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-config\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196474 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196493 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd54a637-b564-42a5-ae67-2daeda3d2f28-tmpfs\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196516 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-apiservice-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196540 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196564 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxp9z\" (UniqueName: \"kubernetes.io/projected/c95c42c9-7795-4e54-8b28-357ad9aa537e-kube-api-access-fxp9z\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196585 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-webhook-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196613 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjj24\" (UniqueName: \"kubernetes.io/projected/89217e6c-9f20-45f0-8a26-7c808efc3359-kube-api-access-fjj24\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196638 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-cabundle\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196661 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.196686 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxbbw\" (UniqueName: \"kubernetes.io/projected/7a8b9b0c-61c8-460d-8226-7794483e1cb1-kube-api-access-kxbbw\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200762 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-service-ca-bundle\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200863 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200899 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200921 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200956 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7a8b9b0c-61c8-460d-8226-7794483e1cb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.200986 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4rvf\" (UniqueName: \"kubernetes.io/projected/cd54a637-b564-42a5-ae67-2daeda3d2f28-kube-api-access-z4rvf\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201030 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-default-certificate\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201058 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201090 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201133 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klx7g\" (UniqueName: \"kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201158 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-profile-collector-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201181 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c95c42c9-7795-4e54-8b28-357ad9aa537e-cert\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201204 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe2f13e-8873-45a6-a5c8-b03affdad616-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201235 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z4mq\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201264 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201299 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-stats-auth\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201346 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-serving-cert\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201371 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5104e975-a44a-4068-902e-589d3fce2b99-proxy-tls\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201394 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f24mk\" (UniqueName: \"kubernetes.io/projected/d21a2da8-a31f-40a3-84b5-db9d331e822e-kube-api-access-f24mk\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201421 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201467 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-images\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201494 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-config\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201516 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e61140b-3686-4f1b-b8b3-56878044f9f1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201542 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201585 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21a2da8-a31f-40a3-84b5-db9d331e822e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201609 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e61140b-3686-4f1b-b8b3-56878044f9f1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201682 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9h8s\" (UniqueName: \"kubernetes.io/projected/ada1191a-da37-466d-a7cc-a7d97b9f78d4-kube-api-access-t9h8s\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201729 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvdjc\" (UniqueName: \"kubernetes.io/projected/307c7632-1a40-49ae-96fc-66cf82b9e8eb-kube-api-access-rvdjc\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201758 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201786 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78175b85-3b98-40b5-aa1a-abc5804132e9-config\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201848 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/89217e6c-9f20-45f0-8a26-7c808efc3359-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201877 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201906 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2wl\" (UniqueName: \"kubernetes.io/projected/5bc4e7d4-84ac-4ab8-ba99-143340209459-kube-api-access-5s2wl\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.201985 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-client\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202041 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-metrics-certs\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202111 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe2f13e-8873-45a6-a5c8-b03affdad616-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202160 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-srv-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202195 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82bd8\" (UniqueName: \"kubernetes.io/projected/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-kube-api-access-82bd8\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202236 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zf58\" (UniqueName: \"kubernetes.io/projected/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-kube-api-access-7zf58\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.202265 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:27.702242441 +0000 UTC m=+137.271460071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202309 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn474\" (UniqueName: \"kubernetes.io/projected/f3125b89-92ef-4575-be6c-fef0fd06e019-kube-api-access-tn474\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202357 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78175b85-3b98-40b5-aa1a-abc5804132e9-serving-cert\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202389 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-node-bootstrap-token\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202409 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-srv-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202431 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctnf8\" (UniqueName: \"kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202454 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rgwm\" (UniqueName: \"kubernetes.io/projected/8671ff24-9cdd-4013-9e9e-f885c5029f58-kube-api-access-5rgwm\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202474 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-certs\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202493 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftpqh\" (UniqueName: \"kubernetes.io/projected/5104e975-a44a-4068-902e-589d3fce2b99-kube-api-access-ftpqh\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202512 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e61140b-3686-4f1b-b8b3-56878044f9f1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.202524 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.203366 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-service-ca-bundle\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.204907 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.208203 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78175b85-3b98-40b5-aa1a-abc5804132e9-config\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.216214 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.219749 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5104e975-a44a-4068-902e-589d3fce2b99-images\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.220624 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe2f13e-8873-45a6-a5c8-b03affdad616-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.227123 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-config\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.227656 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.228459 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-config\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.234074 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-key\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.234638 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cd54a637-b564-42a5-ae67-2daeda3d2f28-tmpfs\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.269201 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.269534 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7a8b9b0c-61c8-460d-8226-7794483e1cb1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.272160 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-profile-collector-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.273557 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e61140b-3686-4f1b-b8b3-56878044f9f1-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.273999 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.274412 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-client\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.275082 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.276275 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.280836 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c95c42c9-7795-4e54-8b28-357ad9aa537e-cert\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.281245 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-default-certificate\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.281784 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.282531 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5bc4e7d4-84ac-4ab8-ba99-143340209459-srv-cert\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.282987 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-webhook-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.284139 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/160c8f8d-4bee-4f69-9051-ee22f8014444-serving-cert\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.284568 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.284870 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.284956 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ada1191a-da37-466d-a7cc-a7d97b9f78d4-srv-cert\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.285395 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/89217e6c-9f20-45f0-8a26-7c808efc3359-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.285487 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe2f13e-8873-45a6-a5c8-b03affdad616-config\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.286012 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-metrics-certs\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.286187 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/160c8f8d-4bee-4f69-9051-ee22f8014444-etcd-service-ca\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.297275 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.298949 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e61140b-3686-4f1b-b8b3-56878044f9f1-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.299549 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/307c7632-1a40-49ae-96fc-66cf82b9e8eb-signing-cabundle\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.310814 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.313143 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-stats-auth\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.313461 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.316244 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9h8s\" (UniqueName: \"kubernetes.io/projected/ada1191a-da37-466d-a7cc-a7d97b9f78d4-kube-api-access-t9h8s\") pod \"olm-operator-6b444d44fb-qvxlv\" (UID: \"ada1191a-da37-466d-a7cc-a7d97b9f78d4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.314801 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.321071 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn474\" (UniqueName: \"kubernetes.io/projected/f3125b89-92ef-4575-be6c-fef0fd06e019-kube-api-access-tn474\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.321145 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-node-bootstrap-token\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.321196 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-certs\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.323188 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78175b85-3b98-40b5-aa1a-abc5804132e9-serving-cert\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.330553 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:27.830533049 +0000 UTC m=+137.399750679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.337743 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cd54a637-b564-42a5-ae67-2daeda3d2f28-apiservice-cert\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.338959 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d21a2da8-a31f-40a3-84b5-db9d331e822e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.350204 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5104e975-a44a-4068-902e-589d3fce2b99-proxy-tls\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.363680 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.363725 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjj24\" (UniqueName: \"kubernetes.io/projected/89217e6c-9f20-45f0-8a26-7c808efc3359-kube-api-access-fjj24\") pod \"control-plane-machine-set-operator-78cbb6b69f-wx4gj\" (UID: \"89217e6c-9f20-45f0-8a26-7c808efc3359\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.364212 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7380efd7-ce34-4ab1-990a-1a0ba3d6f714-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bc62n\" (UID: \"7380efd7-ce34-4ab1-990a-1a0ba3d6f714\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.364256 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z4mq\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.364915 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-certs\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.365233 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvdjc\" (UniqueName: \"kubernetes.io/projected/307c7632-1a40-49ae-96fc-66cf82b9e8eb-kube-api-access-rvdjc\") pod \"service-ca-9c57cc56f-g485q\" (UID: \"307c7632-1a40-49ae-96fc-66cf82b9e8eb\") " pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.365927 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8671ff24-9cdd-4013-9e9e-f885c5029f58-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.366538 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2wl\" (UniqueName: \"kubernetes.io/projected/5bc4e7d4-84ac-4ab8-ba99-143340209459-kube-api-access-5s2wl\") pod \"catalog-operator-68c6474976-52rdk\" (UID: \"5bc4e7d4-84ac-4ab8-ba99-143340209459\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.367214 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3125b89-92ef-4575-be6c-fef0fd06e019-node-bootstrap-token\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.369248 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctnf8\" (UniqueName: \"kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8\") pod \"collect-profiles-29422935-68ckk\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.421022 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" event={"ID":"743fcc8f-13eb-4ed8-a811-25ea265076bb","Type":"ContainerStarted","Data":"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.421345 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" event={"ID":"743fcc8f-13eb-4ed8-a811-25ea265076bb","Type":"ContainerStarted","Data":"af1ddd3dccbe89826ca21ca767cef10f16f4b331ed51ec66c8612a6d0feb1aa4"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.422689 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.423031 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.423396 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:27.923383437 +0000 UTC m=+137.492601067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.424673 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4rvf\" (UniqueName: \"kubernetes.io/projected/cd54a637-b564-42a5-ae67-2daeda3d2f28-kube-api-access-z4rvf\") pod \"packageserver-d55dfcdfc-hr7pq\" (UID: \"cd54a637-b564-42a5-ae67-2daeda3d2f28\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.429341 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5fe2f13e-8873-45a6-a5c8-b03affdad616-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rh5mp\" (UID: \"5fe2f13e-8873-45a6-a5c8-b03affdad616\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.435490 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.444572 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" event={"ID":"def56373-569e-4262-8cbb-9a2d880de58d","Type":"ContainerStarted","Data":"bf0504dd4fcd12df8504a74a40bf290d9bcb0bb1c1e94e3ba0a21a0aba7145e2"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.444625 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" event={"ID":"def56373-569e-4262-8cbb-9a2d880de58d","Type":"ContainerStarted","Data":"6a44f36f7d088ddf1959f4a0375280d1accfc20f9957b71c59073383fdc954fd"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.453954 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" event={"ID":"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e","Type":"ContainerStarted","Data":"85cc50bc73e26825ab3e52c6afe0b0910e9f0599949b31397c640d82a2f0702d"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.454001 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" event={"ID":"6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e","Type":"ContainerStarted","Data":"70cfaf333f33bbe29446bd1250c8cc0d8ac6c7db58077a0b4f3bac44eed54351"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.455166 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f24mk\" (UniqueName: \"kubernetes.io/projected/d21a2da8-a31f-40a3-84b5-db9d331e822e-kube-api-access-f24mk\") pod \"cluster-samples-operator-665b6dd947-pcnwq\" (UID: \"d21a2da8-a31f-40a3-84b5-db9d331e822e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.488292 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rgwm\" (UniqueName: \"kubernetes.io/projected/8671ff24-9cdd-4013-9e9e-f885c5029f58-kube-api-access-5rgwm\") pod \"package-server-manager-789f6589d5-jfqtv\" (UID: \"8671ff24-9cdd-4013-9e9e-f885c5029f58\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.491115 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftpqh\" (UniqueName: \"kubernetes.io/projected/5104e975-a44a-4068-902e-589d3fce2b99-kube-api-access-ftpqh\") pod \"machine-config-operator-74547568cd-qj9z5\" (UID: \"5104e975-a44a-4068-902e-589d3fce2b99\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.494976 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2sr6" event={"ID":"24d0c394-4cf4-452f-a117-215657acb8f2","Type":"ContainerStarted","Data":"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.495027 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2sr6" event={"ID":"24d0c394-4cf4-452f-a117-215657acb8f2","Type":"ContainerStarted","Data":"207d1362d360284dbb7f224a9bfd265caf3c850c58b071ef7b4e5a84d072331b"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.504147 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" event={"ID":"4b3fc300-4e72-4e72-9ff2-ad647815fac8","Type":"ContainerStarted","Data":"f81589ec27191aa9d11036d5bf4b27d306132ed314e345f0e8476349c16e2791"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.504205 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" event={"ID":"4b3fc300-4e72-4e72-9ff2-ad647815fac8","Type":"ContainerStarted","Data":"017d1cf81fd5e6458a326ec111aef2dd364f98c779acd85ca72489efc03b269a"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.510309 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" event={"ID":"2e6718b4-c798-4e9a-9dff-fd3bcd9566db","Type":"ContainerStarted","Data":"448d2329f9c2e99923cd8f25a367e6863c758ea2cec9128aab5b265a7fccb01c"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.510388 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" event={"ID":"2e6718b4-c798-4e9a-9dff-fd3bcd9566db","Type":"ContainerStarted","Data":"e62805c05c7a6758b32378a78bd5fcd56cb4b8e34248d8c7128ac6758a96ba0a"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.511947 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.515870 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" event={"ID":"57fb36d2-42fe-4993-a454-1db53581dc35","Type":"ContainerStarted","Data":"769a42ca0bd4fc14bac7721210cdca59a1586a68d87b773caff3da176dfde587"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.515931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" event={"ID":"57fb36d2-42fe-4993-a454-1db53581dc35","Type":"ContainerStarted","Data":"d277816c7c9112293ecd7126ec6899fae1407ba2ac42f7383db65e4944e162c9"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.515945 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" event={"ID":"57fb36d2-42fe-4993-a454-1db53581dc35","Type":"ContainerStarted","Data":"f787f597483142b23de0a6cbed38e22712c6ba608134351aed1a03709e901517"} Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.522215 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e61140b-3686-4f1b-b8b3-56878044f9f1-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x7x5p\" (UID: \"1e61140b-3686-4f1b-b8b3-56878044f9f1\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.522359 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.524317 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.528605 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.02858412 +0000 UTC m=+137.597801750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.529328 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.530374 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.535306 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.543282 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.556611 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxp9z\" (UniqueName: \"kubernetes.io/projected/c95c42c9-7795-4e54-8b28-357ad9aa537e-kube-api-access-fxp9z\") pod \"ingress-canary-mzz74\" (UID: \"c95c42c9-7795-4e54-8b28-357ad9aa537e\") " pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.558423 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.565610 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.572135 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82bd8\" (UniqueName: \"kubernetes.io/projected/2af0a9e4-7df0-470f-b06b-ea78cf5b085f-kube-api-access-82bd8\") pod \"kube-storage-version-migrator-operator-b67b599dd-6wrm5\" (UID: \"2af0a9e4-7df0-470f-b06b-ea78cf5b085f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.584788 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.599367 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.602614 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.605585 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zf58\" (UniqueName: \"kubernetes.io/projected/1d38cc32-7b1d-47f0-bcac-1be811fd4d67-kube-api-access-7zf58\") pod \"router-default-5444994796-78j8d\" (UID: \"1d38cc32-7b1d-47f0-bcac-1be811fd4d67\") " pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.608606 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klx7g\" (UniqueName: \"kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g\") pod \"marketplace-operator-79b997595-98dk2\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.610031 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.616976 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-g485q" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.623986 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.627103 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.628841 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.128816783 +0000 UTC m=+137.698034423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.644013 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.645009 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.667920 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6j76\" (UniqueName: \"kubernetes.io/projected/160c8f8d-4bee-4f69-9051-ee22f8014444-kube-api-access-x6j76\") pod \"etcd-operator-b45778765-pg8lp\" (UID: \"160c8f8d-4bee-4f69-9051-ee22f8014444\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.674193 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4kqn\" (UniqueName: \"kubernetes.io/projected/78175b85-3b98-40b5-aa1a-abc5804132e9-kube-api-access-r4kqn\") pod \"service-ca-operator-777779d784-qfnsr\" (UID: \"78175b85-3b98-40b5-aa1a-abc5804132e9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.674649 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxbbw\" (UniqueName: \"kubernetes.io/projected/7a8b9b0c-61c8-460d-8226-7794483e1cb1-kube-api-access-kxbbw\") pod \"multus-admission-controller-857f4d67dd-5wj5h\" (UID: \"7a8b9b0c-61c8-460d-8226-7794483e1cb1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.678940 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mzz74" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.689967 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn474\" (UniqueName: \"kubernetes.io/projected/f3125b89-92ef-4575-be6c-fef0fd06e019-kube-api-access-tn474\") pod \"machine-config-server-h6hmg\" (UID: \"f3125b89-92ef-4575-be6c-fef0fd06e019\") " pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.691595 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h6hmg" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.732535 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.746869 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.246845084 +0000 UTC m=+137.816062714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.788806 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.802867 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.832220 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb"] Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.844013 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.844729 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.845056 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.345039361 +0000 UTC m=+137.914256991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.858218 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.879946 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.887511 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m2sr6" podStartSLOduration=116.887495349 podStartE2EDuration="1m56.887495349s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:27.887188401 +0000 UTC m=+137.456406031" watchObservedRunningTime="2025-12-10 14:26:27.887495349 +0000 UTC m=+137.456712969" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.888675 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:27 crc kubenswrapper[4847]: I1210 14:26:27.948603 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:27 crc kubenswrapper[4847]: E1210 14:26:27.948973 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.448957749 +0000 UTC m=+138.018175379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: W1210 14:26:27.991255 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1857b2d_62af_42c8_94c7_6fa4c248e521.slice/crio-ec2abc875b02695bbecf7048f8625691eede4862ca77b9a049e5fd6580640b4f WatchSource:0}: Error finding container ec2abc875b02695bbecf7048f8625691eede4862ca77b9a049e5fd6580640b4f: Status 404 returned error can't find the container with id ec2abc875b02695bbecf7048f8625691eede4862ca77b9a049e5fd6580640b4f Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:27.999078 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c6lv4"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.039875 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.056245 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.056621 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.55660598 +0000 UTC m=+138.125823610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.072376 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbdjm"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.158876 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.159292 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.659277392 +0000 UTC m=+138.228495032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.161626 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ppmc4"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.261133 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.261494 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.7614748 +0000 UTC m=+138.330692440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.261668 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.262702 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.762378485 +0000 UTC m=+138.331596115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.311053 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6qqn9"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.315753 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-csskz"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.320814 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.320923 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.354120 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.369616 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.370160 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.87013306 +0000 UTC m=+138.439350690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.389944 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tmhm6" podStartSLOduration=117.389922523 podStartE2EDuration="1m57.389922523s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.358510545 +0000 UTC m=+137.927728175" watchObservedRunningTime="2025-12-10 14:26:28.389922523 +0000 UTC m=+137.959140153" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.408172 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.457924 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.472427 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.472864 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:28.972847453 +0000 UTC m=+138.542065093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.554289 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cmqtl" podStartSLOduration=117.554273331 podStartE2EDuration="1m57.554273331s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.551334519 +0000 UTC m=+138.120552149" watchObservedRunningTime="2025-12-10 14:26:28.554273331 +0000 UTC m=+138.123490961" Dec 10 14:26:28 crc kubenswrapper[4847]: W1210 14:26:28.576496 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8123d07a_88ab_42d4_b554_1a6062fec387.slice/crio-1c5a0deadd0791e8f377455b51e41f314e6018168bbef845d5dbe19414bf2e46 WatchSource:0}: Error finding container 1c5a0deadd0791e8f377455b51e41f314e6018168bbef845d5dbe19414bf2e46: Status 404 returned error can't find the container with id 1c5a0deadd0791e8f377455b51e41f314e6018168bbef845d5dbe19414bf2e46 Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.578840 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.578998 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv"] Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.579167 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.079151677 +0000 UTC m=+138.648369317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: W1210 14:26:28.606353 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fc9d601_18de_40ac_b15d_6d385f61124e.slice/crio-85238fc84ebabb85798c574bcfbaad77606945c2654493e0456e9bf03a4cc053 WatchSource:0}: Error finding container 85238fc84ebabb85798c574bcfbaad77606945c2654493e0456e9bf03a4cc053: Status 404 returned error can't find the container with id 85238fc84ebabb85798c574bcfbaad77606945c2654493e0456e9bf03a4cc053 Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.614815 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj"] Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.628069 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h6hmg" event={"ID":"f3125b89-92ef-4575-be6c-fef0fd06e019","Type":"ContainerStarted","Data":"55ff0b6210a0af27914762312e602fa9b382c858718eb1210cf9b1744d114b53"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.647552 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c6lv4" event={"ID":"10f6f206-7866-45bf-a629-0eb1f832cda4","Type":"ContainerStarted","Data":"b864be826d510542e5c69082e9a5816606b3575d2b0dd37dfd54f82b1370d42d"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.649608 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" podStartSLOduration=117.649587307 podStartE2EDuration="1m57.649587307s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.627853519 +0000 UTC m=+138.197071149" watchObservedRunningTime="2025-12-10 14:26:28.649587307 +0000 UTC m=+138.218804937" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.663263 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-74bhj" podStartSLOduration=117.663243579 podStartE2EDuration="1m57.663243579s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.660515953 +0000 UTC m=+138.229733583" watchObservedRunningTime="2025-12-10 14:26:28.663243579 +0000 UTC m=+138.232461209" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.680127 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" event={"ID":"be8157de-bba2-44f2-b301-031caf2f46f5","Type":"ContainerStarted","Data":"33541cc7f6d26035dc83cc6b62dfc2b2a2f47e5de5dc1d56a2d17e8d124961bd"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.680813 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.681195 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.18116045 +0000 UTC m=+138.750378080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.697117 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" event={"ID":"c1857b2d-62af-42c8-94c7-6fa4c248e521","Type":"ContainerStarted","Data":"ec2abc875b02695bbecf7048f8625691eede4862ca77b9a049e5fd6580640b4f"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.700733 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" event={"ID":"a0fed061-99e2-4dc2-98cb-a81937b7d7b8","Type":"ContainerStarted","Data":"07d7bc7e3f2b3cc77a2a940f84ea102165fc76399953348568267f38247a38fc"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.709831 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" event={"ID":"9b826590-f030-4a21-a7b2-ac77deda13e5","Type":"ContainerStarted","Data":"5dbe13931f0a8d5eed05fa1de9ec52a83ce9dbf9caf1891b20d52d82f5ffee33"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.715486 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6qqn9" event={"ID":"420a02dc-3c91-4c2c-b885-06fa71cb2e17","Type":"ContainerStarted","Data":"4f5aed7c0af0211115c9c92f3e6ad3d2ff0d400c8d5d830174dd4c0099e30574"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.733403 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" event={"ID":"176f2b83-e66f-43af-8354-69f2ee03315a","Type":"ContainerStarted","Data":"8100089b7f0ff5405c81acc627b87b3ebbe05f323eff9507d6d6110548231414"} Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.782558 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.782690 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.282661839 +0000 UTC m=+138.851879469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.783179 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.786645 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.28662354 +0000 UTC m=+138.855841240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.860284 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-79pvc" podStartSLOduration=117.86026161 podStartE2EDuration="1m57.86026161s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.860046304 +0000 UTC m=+138.429263934" watchObservedRunningTime="2025-12-10 14:26:28.86026161 +0000 UTC m=+138.429479240" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.884167 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.885996 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.385971099 +0000 UTC m=+138.955188739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.897230 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" podStartSLOduration=117.897202623 podStartE2EDuration="1m57.897202623s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:28.89533095 +0000 UTC m=+138.464548580" watchObservedRunningTime="2025-12-10 14:26:28.897202623 +0000 UTC m=+138.466420253" Dec 10 14:26:28 crc kubenswrapper[4847]: I1210 14:26:28.987806 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:28 crc kubenswrapper[4847]: E1210 14:26:28.988229 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.488212399 +0000 UTC m=+139.057430029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.071639 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wj5h"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.089034 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.089437 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.58942268 +0000 UTC m=+139.158640310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.089586 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.090000 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.589992106 +0000 UTC m=+139.159209736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.101408 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.123749 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-ml27d" podStartSLOduration=118.123699429 podStartE2EDuration="1m58.123699429s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:29.122266488 +0000 UTC m=+138.691484118" watchObservedRunningTime="2025-12-10 14:26:29.123699429 +0000 UTC m=+138.692917059" Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.192045 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.192382 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.692368389 +0000 UTC m=+139.261586019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.242426 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" podStartSLOduration=118.242406289 podStartE2EDuration="1m58.242406289s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:29.213754748 +0000 UTC m=+138.782972398" watchObservedRunningTime="2025-12-10 14:26:29.242406289 +0000 UTC m=+138.811623919" Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.285169 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cbh" podStartSLOduration=118.285133185 podStartE2EDuration="1m58.285133185s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:29.282098189 +0000 UTC m=+138.851315819" watchObservedRunningTime="2025-12-10 14:26:29.285133185 +0000 UTC m=+138.854350815" Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.294547 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.294909 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.794897128 +0000 UTC m=+139.364114758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: W1210 14:26:29.378395 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89217e6c_9f20_45f0_8a26_7c808efc3359.slice/crio-5d19d4858ca5f36ed466cbe06cf5d344a5fb41b9b198543e47430abcfdfedda9 WatchSource:0}: Error finding container 5d19d4858ca5f36ed466cbe06cf5d344a5fb41b9b198543e47430abcfdfedda9: Status 404 returned error can't find the container with id 5d19d4858ca5f36ed466cbe06cf5d344a5fb41b9b198543e47430abcfdfedda9 Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.386475 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp"] Dec 10 14:26:29 crc kubenswrapper[4847]: W1210 14:26:29.388262 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e61140b_3686_4f1b_b8b3_56878044f9f1.slice/crio-9c444fa1ce7975f85c65baeb442ba4ba3e054d153c1c857c1d9d7b062cb9b1b9 WatchSource:0}: Error finding container 9c444fa1ce7975f85c65baeb442ba4ba3e054d153c1c857c1d9d7b062cb9b1b9: Status 404 returned error can't find the container with id 9c444fa1ce7975f85c65baeb442ba4ba3e054d153c1c857c1d9d7b062cb9b1b9 Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.391912 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-g485q"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.395510 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.396181 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.89616609 +0000 UTC m=+139.465383720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.398827 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mzz74"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.398881 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" podStartSLOduration=118.398871456 podStartE2EDuration="1m58.398871456s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:29.395900783 +0000 UTC m=+138.965118413" watchObservedRunningTime="2025-12-10 14:26:29.398871456 +0000 UTC m=+138.968089086" Dec 10 14:26:29 crc kubenswrapper[4847]: W1210 14:26:29.409939 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a8b9b0c_61c8_460d_8226_7794483e1cb1.slice/crio-0d1b77ed016c80e5e180778afda0a21f9e8beca6d4a89ef31f3b0a1e0bdb71df WatchSource:0}: Error finding container 0d1b77ed016c80e5e180778afda0a21f9e8beca6d4a89ef31f3b0a1e0bdb71df: Status 404 returned error can't find the container with id 0d1b77ed016c80e5e180778afda0a21f9e8beca6d4a89ef31f3b0a1e0bdb71df Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.412780 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.497380 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.497764 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:29.997751012 +0000 UTC m=+139.566968642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.501305 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6lxgl" podStartSLOduration=118.501282041 podStartE2EDuration="1m58.501282041s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:29.450175251 +0000 UTC m=+139.019392881" watchObservedRunningTime="2025-12-10 14:26:29.501282041 +0000 UTC m=+139.070499671" Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.502648 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.531550 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.565777 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.600551 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.601327 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.101303618 +0000 UTC m=+139.670521248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.613245 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.643242 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.704948 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.705440 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.205419551 +0000 UTC m=+139.774637181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.717007 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.720774 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pg8lp"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.729873 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr"] Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.783656 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" event={"ID":"7a8b9b0c-61c8-460d-8226-7794483e1cb1","Type":"ContainerStarted","Data":"0d1b77ed016c80e5e180778afda0a21f9e8beca6d4a89ef31f3b0a1e0bdb71df"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.803592 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" event={"ID":"89217e6c-9f20-45f0-8a26-7c808efc3359","Type":"ContainerStarted","Data":"5d19d4858ca5f36ed466cbe06cf5d344a5fb41b9b198543e47430abcfdfedda9"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.807547 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.808137 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.308117894 +0000 UTC m=+139.877335524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.815019 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" event={"ID":"d21a2da8-a31f-40a3-84b5-db9d331e822e","Type":"ContainerStarted","Data":"1e31db247dbe0b09d117173cb818dbaa426dfa06755c71c3431580c870c5409c"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.822548 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerStarted","Data":"b97d0fb9b31de057f4b3cd61eb4c5909dab0ca9060c308948a2df7bd4828659e"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.833511 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-g485q" event={"ID":"307c7632-1a40-49ae-96fc-66cf82b9e8eb","Type":"ContainerStarted","Data":"0930acbc0a6fda8c9d63554d7baddcd5515529ea9f4189370c0c41c212cd59b7"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.838119 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-78j8d" event={"ID":"1d38cc32-7b1d-47f0-bcac-1be811fd4d67","Type":"ContainerStarted","Data":"24bb6e67f734996fb00b80d5d8cf7321929c9e0db4983c9802f057864f665399"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.845699 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" event={"ID":"9fc9d601-18de-40ac-b15d-6d385f61124e","Type":"ContainerStarted","Data":"85238fc84ebabb85798c574bcfbaad77606945c2654493e0456e9bf03a4cc053"} Dec 10 14:26:29 crc kubenswrapper[4847]: W1210 14:26:29.860866 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bc4e7d4_84ac_4ab8_ba99_143340209459.slice/crio-c41178f3797d1ab20e9be8d7d43e9187484af950d4313075d4a64741b32eaff2 WatchSource:0}: Error finding container c41178f3797d1ab20e9be8d7d43e9187484af950d4313075d4a64741b32eaff2: Status 404 returned error can't find the container with id c41178f3797d1ab20e9be8d7d43e9187484af950d4313075d4a64741b32eaff2 Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.863482 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mzz74" event={"ID":"c95c42c9-7795-4e54-8b28-357ad9aa537e","Type":"ContainerStarted","Data":"ccaa4d7af2744c7a5c08aa6adcac77b290968ec4a0ef3f6a33f49437c7ccd1f1"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.876966 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" event={"ID":"cd54a637-b564-42a5-ae67-2daeda3d2f28","Type":"ContainerStarted","Data":"e1eacd52ff76924921278fe71c769b34b91367e6f96450c1bd77c720644bd6e7"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.882465 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" event={"ID":"ada1191a-da37-466d-a7cc-a7d97b9f78d4","Type":"ContainerStarted","Data":"6451fde13b3abd124108fdb510d2e9e1715ebe02671ff64ad62845f8147be881"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.887904 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" event={"ID":"5fe2f13e-8873-45a6-a5c8-b03affdad616","Type":"ContainerStarted","Data":"54d9ceefae0b15b5b80b51199818d03b36d06bc28c7d0f6fbfcdb866cc978dc3"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.893082 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" event={"ID":"7380efd7-ce34-4ab1-990a-1a0ba3d6f714","Type":"ContainerStarted","Data":"021f79c4017f13a572e05a00b15e1e338ecc1a045c56aa93082152903d400b72"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.901219 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" event={"ID":"5104e975-a44a-4068-902e-589d3fce2b99","Type":"ContainerStarted","Data":"27b386af4385bb094f39d6d1a9c294b9694ba7a5232012866a3953f12e75fd4a"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.907536 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" event={"ID":"69b03596-eace-45ce-ad89-bd52e7fd5766","Type":"ContainerStarted","Data":"cc94be4913fc543c243501974fba1e5a2ac397174abb52ce2e353c01993d7a8d"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.908872 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:29 crc kubenswrapper[4847]: E1210 14:26:29.909317 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.409302564 +0000 UTC m=+139.978520194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.934984 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" event={"ID":"8123d07a-88ab-42d4-b554-1a6062fec387","Type":"ContainerStarted","Data":"1c5a0deadd0791e8f377455b51e41f314e6018168bbef845d5dbe19414bf2e46"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.941507 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" event={"ID":"8671ff24-9cdd-4013-9e9e-f885c5029f58","Type":"ContainerStarted","Data":"1d144ab103f77c3fea6ef50b3e60c55001192ef4c5b96fb448af0bf36978240b"} Dec 10 14:26:29 crc kubenswrapper[4847]: I1210 14:26:29.950507 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" event={"ID":"1e61140b-3686-4f1b-b8b3-56878044f9f1","Type":"ContainerStarted","Data":"9c444fa1ce7975f85c65baeb442ba4ba3e054d153c1c857c1d9d7b062cb9b1b9"} Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.013298 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.014807 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.514779395 +0000 UTC m=+140.083997025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.115016 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.115669 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.615654466 +0000 UTC m=+140.184872096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.219093 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.219543 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.719523602 +0000 UTC m=+140.288741222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.322359 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.322842 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.822825742 +0000 UTC m=+140.392043362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.369585 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.370011 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.399548 4847 patch_prober.go:28] interesting pod/apiserver-76f77b778f-jvtvd container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]log ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]etcd ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/max-in-flight-filter ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 10 14:26:30 crc kubenswrapper[4847]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 10 14:26:30 crc kubenswrapper[4847]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/project.openshift.io-projectcache ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-startinformers ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 10 14:26:30 crc kubenswrapper[4847]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 14:26:30 crc kubenswrapper[4847]: livez check failed Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.399615 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" podUID="6b873c5d-e3ab-41b1-a8ba-10aa5b446c9e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.424110 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.424388 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.924349081 +0000 UTC m=+140.493566721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.424470 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.429536 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:30.929516476 +0000 UTC m=+140.498734186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.525935 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.527805 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.027775224 +0000 UTC m=+140.596992864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.628361 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.629076 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.129062047 +0000 UTC m=+140.698279677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.732926 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.733115 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.233087907 +0000 UTC m=+140.802305537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.733205 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.733569 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.233557081 +0000 UTC m=+140.802774711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.841268 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.842123 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.342100217 +0000 UTC m=+140.911317847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:30 crc kubenswrapper[4847]: I1210 14:26:30.943216 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:30 crc kubenswrapper[4847]: E1210 14:26:30.943556 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.443542955 +0000 UTC m=+141.012760585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.038385 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" event={"ID":"8671ff24-9cdd-4013-9e9e-f885c5029f58","Type":"ContainerStarted","Data":"7488b0bae407109cfd43aab7fcd9d01fd492758e38ac9457f9844b6ca1e1301e"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.045871 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.046215 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.546200286 +0000 UTC m=+141.115417916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.048433 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" event={"ID":"160c8f8d-4bee-4f69-9051-ee22f8014444","Type":"ContainerStarted","Data":"f3c1eae2492104ec5933bab91e9d69cb56116384f7f4e8bbec82617eb724501a"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.048473 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" event={"ID":"160c8f8d-4bee-4f69-9051-ee22f8014444","Type":"ContainerStarted","Data":"a925ff971accb15a00b55963eb31aa8ff36329a0033c2a86ea7f1190f2488d3a"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.075356 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mzz74" event={"ID":"c95c42c9-7795-4e54-8b28-357ad9aa537e","Type":"ContainerStarted","Data":"d3e2ab5ef762cf774d81860be424a784549f3bf922290623b4d011bd073c45bd"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.086882 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-78j8d" event={"ID":"1d38cc32-7b1d-47f0-bcac-1be811fd4d67","Type":"ContainerStarted","Data":"35915fc801a0deca9b410ccd63b2501e560a9dab0151d1327970410dadcf77f1"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.150051 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.151778 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerStarted","Data":"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.152704 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.160129 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.660110322 +0000 UTC m=+141.229327952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.177985 4847 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-98dk2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.178373 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.179211 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" event={"ID":"cd54a637-b564-42a5-ae67-2daeda3d2f28","Type":"ContainerStarted","Data":"0dbc4e3fe0e3835163594b9113aa58853276f4af0eb79025024c983cf186a20c"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.180062 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.202596 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" event={"ID":"8123d07a-88ab-42d4-b554-1a6062fec387","Type":"ContainerStarted","Data":"a131a86dfab102769afa283d8285bd45eb8ff4849fd46c192738bb96b43397e9"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.238454 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" event={"ID":"176f2b83-e66f-43af-8354-69f2ee03315a","Type":"ContainerStarted","Data":"969fd4712894ba1df82eb1dc3be0de517bcc270be659abcd69bb263a756e67ae"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.250677 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.251727 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.751671804 +0000 UTC m=+141.320889434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.260861 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" event={"ID":"5bc4e7d4-84ac-4ab8-ba99-143340209459","Type":"ContainerStarted","Data":"148e2d208a173f8995d4ba10dc7a4598bef1d4be8e112df784d3916c4d405817"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.260899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" event={"ID":"5bc4e7d4-84ac-4ab8-ba99-143340209459","Type":"ContainerStarted","Data":"c41178f3797d1ab20e9be8d7d43e9187484af950d4313075d4a64741b32eaff2"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.261818 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.285486 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" event={"ID":"be8157de-bba2-44f2-b301-031caf2f46f5","Type":"ContainerStarted","Data":"9ed91d49ea6622734e5107d208f9bd917b963a4f2bc5a17d243e677ca08a70c9"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.319883 4847 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-52rdk container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.319961 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" podUID="5bc4e7d4-84ac-4ab8-ba99-143340209459" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.328919 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" event={"ID":"7380efd7-ce34-4ab1-990a-1a0ba3d6f714","Type":"ContainerStarted","Data":"948a5350eb61725653a3eca55d37a09fb1b750a782d2e7287b9c0f7cf96c264c"} Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.361124 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.861109085 +0000 UTC m=+141.430326715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.366064 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" event={"ID":"2af0a9e4-7df0-470f-b06b-ea78cf5b085f","Type":"ContainerStarted","Data":"447ab6e6a9e65140bf9e77a649d593e5b89b35ad6a6d52c13a1e29998c73e7e6"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.352164 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.422162 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" event={"ID":"89217e6c-9f20-45f0-8a26-7c808efc3359","Type":"ContainerStarted","Data":"6cc5849bf740a15c358e52e86d3b529873016701e6752239250fff30a7e7c0d2"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.447451 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" event={"ID":"78175b85-3b98-40b5-aa1a-abc5804132e9","Type":"ContainerStarted","Data":"40c49dae01af4356297eab44907d46364cdb637a27f3e3bdac991779fe25c2e0"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.447505 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" event={"ID":"78175b85-3b98-40b5-aa1a-abc5804132e9","Type":"ContainerStarted","Data":"7b60f276b5737b99042aec1687f3396566260700e797e35ebcaec05b05cdca95"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.471172 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" event={"ID":"a0fed061-99e2-4dc2-98cb-a81937b7d7b8","Type":"ContainerStarted","Data":"21b73fa04e7f7e54b2295951a89b3a9240fbec498151214bc8777652004c4718"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.474744 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.475525 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:31.975494784 +0000 UTC m=+141.544712414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.484208 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" podStartSLOduration=120.484184688 podStartE2EDuration="2m0.484184688s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.472532671 +0000 UTC m=+141.041750321" watchObservedRunningTime="2025-12-10 14:26:31.484184688 +0000 UTC m=+141.053402318" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.508927 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c6lv4" event={"ID":"10f6f206-7866-45bf-a629-0eb1f832cda4","Type":"ContainerStarted","Data":"17e4ff940c19a4cdba8944367764a2d6f381635a75043900c6265f0025501e48"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.509809 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.510911 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qfnsr" podStartSLOduration=120.510891705 podStartE2EDuration="2m0.510891705s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.50999009 +0000 UTC m=+141.079207720" watchObservedRunningTime="2025-12-10 14:26:31.510891705 +0000 UTC m=+141.080109335" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.538500 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6qqn9" event={"ID":"420a02dc-3c91-4c2c-b885-06fa71cb2e17","Type":"ContainerStarted","Data":"69b08e7ad353e1aa8f3043e4e7f430064a3f669f7474fcfd01083a220021ec41"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.538650 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.562097 4847 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6lv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.562158 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6lv4" podUID="10f6f206-7866-45bf-a629-0eb1f832cda4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.575328 4847 generic.go:334] "Generic (PLEG): container finished" podID="c1857b2d-62af-42c8-94c7-6fa4c248e521" containerID="26deba77009f7f011c0d6570ff5f125326a6caba832534dbd5cf015aa961ac2f" exitCode=0 Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.575848 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" event={"ID":"c1857b2d-62af-42c8-94c7-6fa4c248e521","Type":"ContainerDied","Data":"26deba77009f7f011c0d6570ff5f125326a6caba832534dbd5cf015aa961ac2f"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.575974 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.577433 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.077413446 +0000 UTC m=+141.646631156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.593990 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" podStartSLOduration=120.593969069 podStartE2EDuration="2m0.593969069s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.593236528 +0000 UTC m=+141.162454158" watchObservedRunningTime="2025-12-10 14:26:31.593969069 +0000 UTC m=+141.163186699" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.595448 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bc62n" podStartSLOduration=120.595441759 podStartE2EDuration="2m0.595441759s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.547077747 +0000 UTC m=+141.116295377" watchObservedRunningTime="2025-12-10 14:26:31.595441759 +0000 UTC m=+141.164659389" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.608082 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" event={"ID":"5104e975-a44a-4068-902e-589d3fce2b99","Type":"ContainerStarted","Data":"d7f17c95f13cf09af851f41679626cbfb6cd699f36be83ed2c1023fd5a184294"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.621915 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wx4gj" podStartSLOduration=120.62190046 podStartE2EDuration="2m0.62190046s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.619249306 +0000 UTC m=+141.188466936" watchObservedRunningTime="2025-12-10 14:26:31.62190046 +0000 UTC m=+141.191118090" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.647955 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" podStartSLOduration=120.647938778 podStartE2EDuration="2m0.647938778s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.646587951 +0000 UTC m=+141.215805581" watchObservedRunningTime="2025-12-10 14:26:31.647938778 +0000 UTC m=+141.217156408" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.671482 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" event={"ID":"9fc9d601-18de-40ac-b15d-6d385f61124e","Type":"ContainerStarted","Data":"5381d50686b00021c256ea8845a08d61fbbad7d817e054505d36f30d08743e69"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.682543 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.683945 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.183918715 +0000 UTC m=+141.753136345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.698047 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h6hmg" event={"ID":"f3125b89-92ef-4575-be6c-fef0fd06e019","Type":"ContainerStarted","Data":"b0dbfa70ca5aeb3a18fa0c39e81cbbf42b70acd93604f34f906a6e9c81887e8f"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.723008 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" podStartSLOduration=120.722989148 podStartE2EDuration="2m0.722989148s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.681808625 +0000 UTC m=+141.251026255" watchObservedRunningTime="2025-12-10 14:26:31.722989148 +0000 UTC m=+141.292206778" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.723899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-g485q" event={"ID":"307c7632-1a40-49ae-96fc-66cf82b9e8eb","Type":"ContainerStarted","Data":"3600cf38cff4a09aa7f7ea4c10b6edc4717e0aaee2bd37d98e99b14e7e9d6553"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.752425 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" event={"ID":"69b03596-eace-45ce-ad89-bd52e7fd5766","Type":"ContainerStarted","Data":"5472ceb72d8aade48384de41aa34f010fe3b922e7059dce135fe308cd4d21621"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.752917 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" podStartSLOduration=120.752894344 podStartE2EDuration="2m0.752894344s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.747491833 +0000 UTC m=+141.316709463" watchObservedRunningTime="2025-12-10 14:26:31.752894344 +0000 UTC m=+141.322111974" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.753740 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-pg8lp" podStartSLOduration=120.753734658 podStartE2EDuration="2m0.753734658s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.724905541 +0000 UTC m=+141.294123171" watchObservedRunningTime="2025-12-10 14:26:31.753734658 +0000 UTC m=+141.322952308" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.762969 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" event={"ID":"ada1191a-da37-466d-a7cc-a7d97b9f78d4","Type":"ContainerStarted","Data":"215f3fb484f103cad339b620d7007ca130c93c1226ca6bc23b646f74ae69510d"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.763806 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.774778 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" event={"ID":"d21a2da8-a31f-40a3-84b5-db9d331e822e","Type":"ContainerStarted","Data":"da060674cc453e1bf230796ab7551c2f212f5aa5acbe6b3db999bcad49461d13"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.790407 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.793309 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-78j8d" podStartSLOduration=120.793291414 podStartE2EDuration="2m0.793291414s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.789534809 +0000 UTC m=+141.358752449" watchObservedRunningTime="2025-12-10 14:26:31.793291414 +0000 UTC m=+141.362509044" Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.793663 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.293646124 +0000 UTC m=+141.862863744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.806254 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b826590-f030-4a21-a7b2-ac77deda13e5" containerID="71479419160f88f6a5157555832154d955dc07a3c6fb31ea8397a1ca52e24874" exitCode=0 Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.806333 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" event={"ID":"9b826590-f030-4a21-a7b2-ac77deda13e5","Type":"ContainerDied","Data":"71479419160f88f6a5157555832154d955dc07a3c6fb31ea8397a1ca52e24874"} Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.807157 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.828315 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-mzz74" podStartSLOduration=7.828293563 podStartE2EDuration="7.828293563s" podCreationTimestamp="2025-12-10 14:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.817211993 +0000 UTC m=+141.386429643" watchObservedRunningTime="2025-12-10 14:26:31.828293563 +0000 UTC m=+141.397511193" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.853013 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" podStartSLOduration=120.852995254 podStartE2EDuration="2m0.852995254s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.851140123 +0000 UTC m=+141.420357773" watchObservedRunningTime="2025-12-10 14:26:31.852995254 +0000 UTC m=+141.422212884" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.865343 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.877132 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:31 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:31 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:31 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.877197 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.887610 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-g485q" podStartSLOduration=120.887585212 podStartE2EDuration="2m0.887585212s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.882312995 +0000 UTC m=+141.451530625" watchObservedRunningTime="2025-12-10 14:26:31.887585212 +0000 UTC m=+141.456802842" Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.891552 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:31 crc kubenswrapper[4847]: E1210 14:26:31.892591 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.392573771 +0000 UTC m=+141.961791401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:31 crc kubenswrapper[4847]: I1210 14:26:31.947614 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" podStartSLOduration=120.94759657 podStartE2EDuration="2m0.94759657s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.915011419 +0000 UTC m=+141.484229049" watchObservedRunningTime="2025-12-10 14:26:31.94759657 +0000 UTC m=+141.516814200" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.013525 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.013936 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.513915416 +0000 UTC m=+142.083133056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.030654 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" podStartSLOduration=121.030632054 podStartE2EDuration="2m1.030632054s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:31.975138141 +0000 UTC m=+141.544355781" watchObservedRunningTime="2025-12-10 14:26:32.030632054 +0000 UTC m=+141.599849684" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.033255 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" podStartSLOduration=121.033229166 podStartE2EDuration="2m1.033229166s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.029836321 +0000 UTC m=+141.599053971" watchObservedRunningTime="2025-12-10 14:26:32.033229166 +0000 UTC m=+141.602446796" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.070359 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6qqn9" podStartSLOduration=8.070339734 podStartE2EDuration="8.070339734s" podCreationTimestamp="2025-12-10 14:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.067754862 +0000 UTC m=+141.636972512" watchObservedRunningTime="2025-12-10 14:26:32.070339734 +0000 UTC m=+141.639557364" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.096674 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-h6hmg" podStartSLOduration=8.09665139 podStartE2EDuration="8.09665139s" podCreationTimestamp="2025-12-10 14:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.090833917 +0000 UTC m=+141.660051567" watchObservedRunningTime="2025-12-10 14:26:32.09665139 +0000 UTC m=+141.665869020" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.117519 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.123059 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.623040398 +0000 UTC m=+142.192258028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.124221 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-c6lv4" podStartSLOduration=121.12417858 podStartE2EDuration="2m1.12417858s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.124036186 +0000 UTC m=+141.693253816" watchObservedRunningTime="2025-12-10 14:26:32.12417858 +0000 UTC m=+141.693396210" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.162341 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" podStartSLOduration=121.162326437 podStartE2EDuration="2m1.162326437s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.155099925 +0000 UTC m=+141.724317555" watchObservedRunningTime="2025-12-10 14:26:32.162326437 +0000 UTC m=+141.731544067" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.180157 4847 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-hr7pq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.180260 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" podUID="cd54a637-b564-42a5-ae67-2daeda3d2f28" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.188010 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-qvxlv" podStartSLOduration=121.187994945 podStartE2EDuration="2m1.187994945s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.185737012 +0000 UTC m=+141.754954642" watchObservedRunningTime="2025-12-10 14:26:32.187994945 +0000 UTC m=+141.757212575" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.223806 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.224197 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.724186148 +0000 UTC m=+142.293403778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.324808 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.325464 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.8254482 +0000 UTC m=+142.394665830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.426772 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.427223 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:32.927191066 +0000 UTC m=+142.496408696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.528354 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.528697 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.028683044 +0000 UTC m=+142.597900674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.629518 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.629868 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.129856575 +0000 UTC m=+142.699074205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.648109 4847 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.730741 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.731128 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.231113078 +0000 UTC m=+142.800330708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.832466 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.832968 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f8dtg" event={"ID":"176f2b83-e66f-43af-8354-69f2ee03315a","Type":"ContainerStarted","Data":"b9c209de6a60ed8ee9721dfc901909aa7c52618117502a0bce74457997a65acc"} Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.833187 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.333150511 +0000 UTC m=+142.902368171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.839856 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6qqn9" event={"ID":"420a02dc-3c91-4c2c-b885-06fa71cb2e17","Type":"ContainerStarted","Data":"541bb0b004929f67b871645db061c204d887cc93594f46201c399a7346da8ab1"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.856116 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" event={"ID":"c1857b2d-62af-42c8-94c7-6fa4c248e521","Type":"ContainerStarted","Data":"cd64c6c73af65fc8dcf8132bcd7017ec2e23a7bd60b2138052c29c97430266b8"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.859960 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" event={"ID":"d21a2da8-a31f-40a3-84b5-db9d331e822e","Type":"ContainerStarted","Data":"962d6b6689a97d4c51187a4f1c408b8564c6dec8c6f4796782a704f38a3c7b79"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.864198 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:32 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:32 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:32 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.864245 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.871098 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-t95rl" event={"ID":"9fc9d601-18de-40ac-b15d-6d385f61124e","Type":"ContainerStarted","Data":"c2565eb052241fe8785658771d21dc16e57a918c5e264b62067f0fbfa4bfb737"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.875736 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" event={"ID":"7a8b9b0c-61c8-460d-8226-7794483e1cb1","Type":"ContainerStarted","Data":"6ea126602b987900b4186e3cf99c8166a9c57cf790668814f29f537236ace22d"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.875824 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" event={"ID":"7a8b9b0c-61c8-460d-8226-7794483e1cb1","Type":"ContainerStarted","Data":"628d626070a608e945aa403f1bcbc99cee28b8713c8575eada9f58d1488a2178"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.880318 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" event={"ID":"8123d07a-88ab-42d4-b554-1a6062fec387","Type":"ContainerStarted","Data":"d24fc2ec237f9142e3795aa1c781202d41f7ca0847984a9e3babe261a0f91768"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.883088 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" podStartSLOduration=121.883077988 podStartE2EDuration="2m1.883077988s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.882954515 +0000 UTC m=+142.452172155" watchObservedRunningTime="2025-12-10 14:26:32.883077988 +0000 UTC m=+142.452295618" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.885621 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" event={"ID":"5fe2f13e-8873-45a6-a5c8-b03affdad616","Type":"ContainerStarted","Data":"42be5232f81f66cbbde938234b8308ecedb214ee4195921979f1b4fca1987535"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.904818 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qj9z5" event={"ID":"5104e975-a44a-4068-902e-589d3fce2b99","Type":"ContainerStarted","Data":"2e6d2685ad41af98fd471e90f2889d6bb26b4f851cb16d4b2773d058a95a5827"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.913760 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6wrm5" event={"ID":"2af0a9e4-7df0-470f-b06b-ea78cf5b085f","Type":"ContainerStarted","Data":"e413864fd62ef1b0e9554a4425d8db724a3b82e96354e43e76e936b6571e2bdc"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.927086 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" event={"ID":"be8157de-bba2-44f2-b301-031caf2f46f5","Type":"ContainerStarted","Data":"b8192b271c057d43183b1e827a03d76a22a7f4435e1d69794bb813bf01ce7d86"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.927136 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" event={"ID":"be8157de-bba2-44f2-b301-031caf2f46f5","Type":"ContainerStarted","Data":"bdb78f8acffded0edf11770a3765a90cb158526adf03e96830410ab93070a5ac"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.928902 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" event={"ID":"9b826590-f030-4a21-a7b2-ac77deda13e5","Type":"ContainerStarted","Data":"17fa0b9974b5525a07d074a846c03c1145e3a516930bf969a157c94c6e5c8294"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.929100 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.933444 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.933579 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.43354661 +0000 UTC m=+143.002764250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.933949 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:32 crc kubenswrapper[4847]: E1210 14:26:32.935079 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 14:26:33.435061812 +0000 UTC m=+143.004279562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbhc4" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.946188 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" event={"ID":"8671ff24-9cdd-4013-9e9e-f885c5029f58","Type":"ContainerStarted","Data":"f86279fb74482f8362546b7d11d40f435caec0e2ba5511e7daf0d93cb0aced0b"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.946384 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.948136 4847 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-10T14:26:32.648144176Z","Handler":null,"Name":""} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.961957 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" event={"ID":"a0fed061-99e2-4dc2-98cb-a81937b7d7b8","Type":"ContainerStarted","Data":"168a971c4866f972ac47d53f8961367e55169e0b072c7e9c7a692d860a80c2b5"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.971666 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wj5h" podStartSLOduration=121.971642836 podStartE2EDuration="2m1.971642836s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.916662248 +0000 UTC m=+142.485879878" watchObservedRunningTime="2025-12-10 14:26:32.971642836 +0000 UTC m=+142.540860466" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.982235 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x7x5p" event={"ID":"1e61140b-3686-4f1b-b8b3-56878044f9f1","Type":"ContainerStarted","Data":"e94bf50b89f35682e02f3bad5738d1a5bcbf9bb74baf5ee8e8eb2e2bfb53e2fc"} Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.986878 4847 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-98dk2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.986927 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.986946 4847 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6lv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.987006 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6lv4" podUID="10f6f206-7866-45bf-a629-0eb1f832cda4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.992972 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-52rdk" Dec 10 14:26:32 crc kubenswrapper[4847]: I1210 14:26:32.995429 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-hr7pq" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.004951 4847 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.005002 4847 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.021396 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-s58tq" podStartSLOduration=122.021378236 podStartE2EDuration="2m2.021378236s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:32.972641654 +0000 UTC m=+142.541859304" watchObservedRunningTime="2025-12-10 14:26:33.021378236 +0000 UTC m=+142.590595866" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.039206 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.080297 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ppmc4" podStartSLOduration=122.080278994 podStartE2EDuration="2m2.080278994s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:33.026993554 +0000 UTC m=+142.596211184" watchObservedRunningTime="2025-12-10 14:26:33.080278994 +0000 UTC m=+142.649496624" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.128153 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.135368 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rh5mp" podStartSLOduration=122.135343555 podStartE2EDuration="2m2.135343555s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:33.122246748 +0000 UTC m=+142.691464398" watchObservedRunningTime="2025-12-10 14:26:33.135343555 +0000 UTC m=+142.704561185" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.136258 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" podStartSLOduration=122.13625039 podStartE2EDuration="2m2.13625039s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:33.081338044 +0000 UTC m=+142.650555674" watchObservedRunningTime="2025-12-10 14:26:33.13625039 +0000 UTC m=+142.705468010" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.145789 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.200550 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" podStartSLOduration=122.200533548 podStartE2EDuration="2m2.200533548s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:33.19450647 +0000 UTC m=+142.763724090" watchObservedRunningTime="2025-12-10 14:26:33.200533548 +0000 UTC m=+142.769751178" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.208530 4847 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.208598 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.275223 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbhc4\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.501690 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.629956 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.630939 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.641012 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.653001 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.754449 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.754834 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.754866 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj6vw\" (UniqueName: \"kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.762246 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.830826 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.831994 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.834650 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.845553 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.855547 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.855633 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.855663 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj6vw\" (UniqueName: \"kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.857018 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.857149 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.861573 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:33 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:33 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:33 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.861642 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.881146 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj6vw\" (UniqueName: \"kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw\") pod \"certified-operators-kvfs8\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.956749 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksdhn\" (UniqueName: \"kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.956863 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.956939 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.958707 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.993022 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" event={"ID":"be8157de-bba2-44f2-b301-031caf2f46f5","Type":"ContainerStarted","Data":"b964752e18f44272744a4b0b1098783e13ce839fd585706ce185081a042b22ac"} Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.996153 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" event={"ID":"7ef1bd97-be64-4bee-ae8c-06fd651ac197","Type":"ContainerStarted","Data":"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205"} Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.996204 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" event={"ID":"7ef1bd97-be64-4bee-ae8c-06fd651ac197","Type":"ContainerStarted","Data":"6bc09c624c6b917c218b0bfd9eb41240b4039c1a206e9d2e22d48c134b4f8883"} Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.997613 4847 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6lv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 10 14:26:33 crc kubenswrapper[4847]: I1210 14:26:33.997672 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6lv4" podUID="10f6f206-7866-45bf-a629-0eb1f832cda4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.008697 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.014698 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-fbdjm" podStartSLOduration=10.014682562 podStartE2EDuration="10.014682562s" podCreationTimestamp="2025-12-10 14:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:34.014026064 +0000 UTC m=+143.583243694" watchObservedRunningTime="2025-12-10 14:26:34.014682562 +0000 UTC m=+143.583900192" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.021142 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.022278 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.043211 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.059605 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.059678 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.059734 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksdhn\" (UniqueName: \"kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.060429 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.060666 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.081483 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" podStartSLOduration=123.0814555 podStartE2EDuration="2m3.0814555s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:26:34.047662635 +0000 UTC m=+143.616880265" watchObservedRunningTime="2025-12-10 14:26:34.0814555 +0000 UTC m=+143.650673130" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.111386 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksdhn\" (UniqueName: \"kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn\") pod \"community-operators-sm78p\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.154893 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.160703 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.161381 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.161510 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrk4j\" (UniqueName: \"kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.187055 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.219727 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.220944 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.237001 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.262477 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.262845 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrk4j\" (UniqueName: \"kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.262884 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.263065 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.263398 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.283690 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrk4j\" (UniqueName: \"kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j\") pod \"certified-operators-kvx4r\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.340149 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.353502 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:26:34 crc kubenswrapper[4847]: W1210 14:26:34.361018 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3987f250_b286_4eed_86e0_2c9e6abd9e3f.slice/crio-bbf70442d4795435f6ff4b4a5580a1b71eb5d1d941077a90b96e75f703c91624 WatchSource:0}: Error finding container bbf70442d4795435f6ff4b4a5580a1b71eb5d1d941077a90b96e75f703c91624: Status 404 returned error can't find the container with id bbf70442d4795435f6ff4b4a5580a1b71eb5d1d941077a90b96e75f703c91624 Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.363650 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.363741 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.363805 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxtnv\" (UniqueName: \"kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.464897 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.464965 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxtnv\" (UniqueName: \"kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.465014 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.465373 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.465609 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.502585 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxtnv\" (UniqueName: \"kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv\") pod \"community-operators-q6kb7\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.566048 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.570134 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:26:34 crc kubenswrapper[4847]: W1210 14:26:34.611728 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf681c8b8_c1fb_47a9_91c2_9122d0799eb1.slice/crio-21e448667cfbb3aaab4ba11e9f3f132784f1fab6e0f6d144beeac7169c5244c3 WatchSource:0}: Error finding container 21e448667cfbb3aaab4ba11e9f3f132784f1fab6e0f6d144beeac7169c5244c3: Status 404 returned error can't find the container with id 21e448667cfbb3aaab4ba11e9f3f132784f1fab6e0f6d144beeac7169c5244c3 Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.766886 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.804360 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:26:34 crc kubenswrapper[4847]: W1210 14:26:34.850571 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc714c7b0_d2ab_4efe_a2e6_8eec5d4da9d0.slice/crio-e3c6e07680c90b57ca6913b00d94eded585449c3e8cae10eae1036d015002ddb WatchSource:0}: Error finding container e3c6e07680c90b57ca6913b00d94eded585449c3e8cae10eae1036d015002ddb: Status 404 returned error can't find the container with id e3c6e07680c90b57ca6913b00d94eded585449c3e8cae10eae1036d015002ddb Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.862498 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:34 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:34 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:34 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:34 crc kubenswrapper[4847]: I1210 14:26:34.862553 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.001734 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6f20b79-6bb4-4832-a607-88805371713c" containerID="91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48" exitCode=0 Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.001917 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerDied","Data":"91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.002025 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerStarted","Data":"5594a43853cffe3910c58c15db46c0c81f55e34598d150080cd2c0a429a2dda8"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.012930 4847 generic.go:334] "Generic (PLEG): container finished" podID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerID="ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191" exitCode=0 Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.013015 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerDied","Data":"ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.013077 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerStarted","Data":"21e448667cfbb3aaab4ba11e9f3f132784f1fab6e0f6d144beeac7169c5244c3"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.013281 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.025224 4847 generic.go:334] "Generic (PLEG): container finished" podID="69b03596-eace-45ce-ad89-bd52e7fd5766" containerID="5472ceb72d8aade48384de41aa34f010fe3b922e7059dce135fe308cd4d21621" exitCode=0 Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.025292 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" event={"ID":"69b03596-eace-45ce-ad89-bd52e7fd5766","Type":"ContainerDied","Data":"5472ceb72d8aade48384de41aa34f010fe3b922e7059dce135fe308cd4d21621"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.030021 4847 generic.go:334] "Generic (PLEG): container finished" podID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerID="583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952" exitCode=0 Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.030064 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerDied","Data":"583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.030081 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerStarted","Data":"bbf70442d4795435f6ff4b4a5580a1b71eb5d1d941077a90b96e75f703c91624"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.032707 4847 generic.go:334] "Generic (PLEG): container finished" podID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerID="0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95" exitCode=0 Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.033881 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerDied","Data":"0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.033917 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerStarted","Data":"e3c6e07680c90b57ca6913b00d94eded585449c3e8cae10eae1036d015002ddb"} Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.034801 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.374593 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.379549 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jvtvd" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.619736 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.619781 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.621616 4847 patch_prober.go:28] interesting pod/console-f9d7485db-m2sr6 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.621660 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2sr6" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.856307 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.861475 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.866563 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.872594 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.874702 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:35 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:35 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:35 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:35 crc kubenswrapper[4847]: I1210 14:26:35.874785 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.002448 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j64\" (UniqueName: \"kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.002512 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.002608 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.100216 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.103139 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.103291 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.103365 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j64\" (UniqueName: \"kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.103784 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.104382 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.123829 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j64\" (UniqueName: \"kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64\") pod \"redhat-marketplace-4tsg4\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.179071 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-csskz" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.188674 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.229418 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.230389 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.236771 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.305857 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7mhw\" (UniqueName: \"kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.306304 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.306337 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.322143 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.411707 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume\") pod \"69b03596-eace-45ce-ad89-bd52e7fd5766\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.411881 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctnf8\" (UniqueName: \"kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8\") pod \"69b03596-eace-45ce-ad89-bd52e7fd5766\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.412026 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume\") pod \"69b03596-eace-45ce-ad89-bd52e7fd5766\" (UID: \"69b03596-eace-45ce-ad89-bd52e7fd5766\") " Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.412440 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.412508 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.412601 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7mhw\" (UniqueName: \"kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.416628 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume" (OuterVolumeSpecName: "config-volume") pod "69b03596-eace-45ce-ad89-bd52e7fd5766" (UID: "69b03596-eace-45ce-ad89-bd52e7fd5766"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.417110 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.417208 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.428379 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8" (OuterVolumeSpecName: "kube-api-access-ctnf8") pod "69b03596-eace-45ce-ad89-bd52e7fd5766" (UID: "69b03596-eace-45ce-ad89-bd52e7fd5766"). InnerVolumeSpecName "kube-api-access-ctnf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.434966 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "69b03596-eace-45ce-ad89-bd52e7fd5766" (UID: "69b03596-eace-45ce-ad89-bd52e7fd5766"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.438910 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7mhw\" (UniqueName: \"kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw\") pod \"redhat-marketplace-d69ss\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.515731 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctnf8\" (UniqueName: \"kubernetes.io/projected/69b03596-eace-45ce-ad89-bd52e7fd5766-kube-api-access-ctnf8\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.515763 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69b03596-eace-45ce-ad89-bd52e7fd5766-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.515772 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69b03596-eace-45ce-ad89-bd52e7fd5766-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.618628 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.686444 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 14:26:36 crc kubenswrapper[4847]: E1210 14:26:36.687070 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b03596-eace-45ce-ad89-bd52e7fd5766" containerName="collect-profiles" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.687087 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b03596-eace-45ce-ad89-bd52e7fd5766" containerName="collect-profiles" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.687222 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b03596-eace-45ce-ad89-bd52e7fd5766" containerName="collect-profiles" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.687709 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.690155 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.690590 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.694026 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.719625 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.721132 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.820758 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822320 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822356 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822391 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822431 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.822459 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.825198 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.826035 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.827651 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.829375 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.834550 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.837375 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.838135 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.840453 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.841924 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.842442 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.865792 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:36 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:36 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:36 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.865850 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.871121 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.882656 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.885154 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.928812 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.928868 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.928892 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sq9b\" (UniqueName: \"kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:36 crc kubenswrapper[4847]: I1210 14:26:36.932154 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:26:36 crc kubenswrapper[4847]: W1210 14:26:36.979146 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52e58c9f_f9db_48ac_9055_f32856b4b946.slice/crio-6123b75dfae074abd332cfbc6e9e0ed3e31c71f6eb93f00da4810af44e083eab WatchSource:0}: Error finding container 6123b75dfae074abd332cfbc6e9e0ed3e31c71f6eb93f00da4810af44e083eab: Status 404 returned error can't find the container with id 6123b75dfae074abd332cfbc6e9e0ed3e31c71f6eb93f00da4810af44e083eab Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.015043 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.030754 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.030822 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.030842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sq9b\" (UniqueName: \"kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.031670 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.032038 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.051091 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.052030 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.058866 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.059272 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk" event={"ID":"69b03596-eace-45ce-ad89-bd52e7fd5766","Type":"ContainerDied","Data":"cc94be4913fc543c243501974fba1e5a2ac397174abb52ce2e353c01993d7a8d"} Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.059303 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc94be4913fc543c243501974fba1e5a2ac397174abb52ce2e353c01993d7a8d" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.062889 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.072586 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sq9b\" (UniqueName: \"kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b\") pod \"redhat-operators-8gv4z\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.073341 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerStarted","Data":"6123b75dfae074abd332cfbc6e9e0ed3e31c71f6eb93f00da4810af44e083eab"} Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.083264 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerStarted","Data":"943de1cdf53a1370340d632953bae859d9f2ffe53e2b3c3ec4a1a0055fdc7b83"} Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.134227 4847 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6lv4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.134278 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6lv4" podUID="10f6f206-7866-45bf-a629-0eb1f832cda4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.134579 4847 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6lv4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.134594 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-c6lv4" podUID="10f6f206-7866-45bf-a629-0eb1f832cda4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.184078 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.228773 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.230384 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.245546 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.341818 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5g4l\" (UniqueName: \"kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.341880 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.341905 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.442959 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5g4l\" (UniqueName: \"kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.443023 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.443057 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.443686 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.444094 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.467224 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5g4l\" (UniqueName: \"kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l\") pod \"redhat-operators-kqrfk\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.549225 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 14:26:37 crc kubenswrapper[4847]: W1210 14:26:37.570128 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-d9ad4129d0fd04159c8f50ab852e0375735b39cf3d515a6741122c065e7f5be6 WatchSource:0}: Error finding container d9ad4129d0fd04159c8f50ab852e0375735b39cf3d515a6741122c065e7f5be6: Status 404 returned error can't find the container with id d9ad4129d0fd04159c8f50ab852e0375735b39cf3d515a6741122c065e7f5be6 Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.621992 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.628479 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.858980 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.866098 4847 patch_prober.go:28] interesting pod/router-default-5444994796-78j8d container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 14:26:37 crc kubenswrapper[4847]: [-]has-synced failed: reason withheld Dec 10 14:26:37 crc kubenswrapper[4847]: [+]process-running ok Dec 10 14:26:37 crc kubenswrapper[4847]: healthz check failed Dec 10 14:26:37 crc kubenswrapper[4847]: I1210 14:26:37.866192 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-78j8d" podUID="1d38cc32-7b1d-47f0-bcac-1be811fd4d67" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.108514 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fa4dec14b34656c18514fc81a11f77fcf228daee980e193c744dd5b3c4856f82"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.108949 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f8099b4034a4a122ea91698ed78af1fda2f005ff26374a6408c0d2f92bfd27c4"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.125514 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.135761 4847 generic.go:334] "Generic (PLEG): container finished" podID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerID="48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63" exitCode=0 Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.136256 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerDied","Data":"48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.140894 4847 generic.go:334] "Generic (PLEG): container finished" podID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerID="4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378" exitCode=0 Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.141019 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerDied","Data":"4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.141069 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerStarted","Data":"4533312a682cfceb88fbfd162744b930d56ba1245e8a107912d91c3c1ee638bc"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.147930 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8bf97cc880c90813508a9694112960a18c7810f9b96260df31f1536cb2fcbd12"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.147981 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"52b83e5c4f4db666986822110cf23c842f55381b04ea1d58668bdbc5dcd33e9f"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.149072 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.152941 4847 generic.go:334] "Generic (PLEG): container finished" podID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerID="dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc" exitCode=0 Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.153223 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerDied","Data":"dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.164524 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0f4feea89b06b982a78c348b6fe073816a993939ad634b366b95b48c6dffbba6"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.164583 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d9ad4129d0fd04159c8f50ab852e0375735b39cf3d515a6741122c065e7f5be6"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.172124 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"118f4abc-3653-4852-8f11-074074102527","Type":"ContainerStarted","Data":"ffbb997b754dad228edeeb45b36933a326db556bbdffd75c2205f3224ab28664"} Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.180335 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fjjbb" Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.864563 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:38 crc kubenswrapper[4847]: I1210 14:26:38.870956 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-78j8d" Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.207081 4847 generic.go:334] "Generic (PLEG): container finished" podID="118f4abc-3653-4852-8f11-074074102527" containerID="c53ac356fb17c9797a4d0fd08b1ae1cb050763e5d75e44eb0590aea558010dcd" exitCode=0 Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.207784 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"118f4abc-3653-4852-8f11-074074102527","Type":"ContainerDied","Data":"c53ac356fb17c9797a4d0fd08b1ae1cb050763e5d75e44eb0590aea558010dcd"} Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.216644 4847 generic.go:334] "Generic (PLEG): container finished" podID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerID="f9f3622f1ea9c66bea169cf8e7b965a9959fc8f7457c1e1592f14e662fb63f77" exitCode=0 Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.216706 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerDied","Data":"f9f3622f1ea9c66bea169cf8e7b965a9959fc8f7457c1e1592f14e662fb63f77"} Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.217844 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerStarted","Data":"df1cc3e7362c7cb0db5a66bbe25ac6f8f5faa2c71733c1cd827446180b75ecf5"} Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.816822 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.818137 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.822556 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.823054 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.827073 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.916475 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:39 crc kubenswrapper[4847]: I1210 14:26:39.917664 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.018664 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.019012 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.019112 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.055918 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.145042 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.543742 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 14:26:40 crc kubenswrapper[4847]: W1210 14:26:40.557317 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod98581445_9906_4999_8a55_0e7e6fb218b6.slice/crio-6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4 WatchSource:0}: Error finding container 6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4: Status 404 returned error can't find the container with id 6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4 Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.814563 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.940178 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access\") pod \"118f4abc-3653-4852-8f11-074074102527\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.940257 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir\") pod \"118f4abc-3653-4852-8f11-074074102527\" (UID: \"118f4abc-3653-4852-8f11-074074102527\") " Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.940576 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "118f4abc-3653-4852-8f11-074074102527" (UID: "118f4abc-3653-4852-8f11-074074102527"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:26:40 crc kubenswrapper[4847]: I1210 14:26:40.956156 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "118f4abc-3653-4852-8f11-074074102527" (UID: "118f4abc-3653-4852-8f11-074074102527"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.059937 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/118f4abc-3653-4852-8f11-074074102527-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.059971 4847 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/118f4abc-3653-4852-8f11-074074102527-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.331571 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"118f4abc-3653-4852-8f11-074074102527","Type":"ContainerDied","Data":"ffbb997b754dad228edeeb45b36933a326db556bbdffd75c2205f3224ab28664"} Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.331974 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffbb997b754dad228edeeb45b36933a326db556bbdffd75c2205f3224ab28664" Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.331736 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 14:26:41 crc kubenswrapper[4847]: I1210 14:26:41.363851 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"98581445-9906-4999-8a55-0e7e6fb218b6","Type":"ContainerStarted","Data":"6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4"} Dec 10 14:26:42 crc kubenswrapper[4847]: I1210 14:26:42.101071 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6qqn9" Dec 10 14:26:42 crc kubenswrapper[4847]: I1210 14:26:42.376457 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"98581445-9906-4999-8a55-0e7e6fb218b6","Type":"ContainerStarted","Data":"1dc56eb9e75f5d5b5ce7e5da8d7d92aa6e9edcd6564f564d8a3b5f8b604056d5"} Dec 10 14:26:43 crc kubenswrapper[4847]: I1210 14:26:43.386966 4847 generic.go:334] "Generic (PLEG): container finished" podID="98581445-9906-4999-8a55-0e7e6fb218b6" containerID="1dc56eb9e75f5d5b5ce7e5da8d7d92aa6e9edcd6564f564d8a3b5f8b604056d5" exitCode=0 Dec 10 14:26:43 crc kubenswrapper[4847]: I1210 14:26:43.387010 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"98581445-9906-4999-8a55-0e7e6fb218b6","Type":"ContainerDied","Data":"1dc56eb9e75f5d5b5ce7e5da8d7d92aa6e9edcd6564f564d8a3b5f8b604056d5"} Dec 10 14:26:44 crc kubenswrapper[4847]: I1210 14:26:44.408066 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-pcnwq_d21a2da8-a31f-40a3-84b5-db9d331e822e/cluster-samples-operator/0.log" Dec 10 14:26:44 crc kubenswrapper[4847]: I1210 14:26:44.408206 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" event={"ID":"d21a2da8-a31f-40a3-84b5-db9d331e822e","Type":"ContainerDied","Data":"da060674cc453e1bf230796ab7551c2f212f5aa5acbe6b3db999bcad49461d13"} Dec 10 14:26:44 crc kubenswrapper[4847]: I1210 14:26:44.408789 4847 generic.go:334] "Generic (PLEG): container finished" podID="d21a2da8-a31f-40a3-84b5-db9d331e822e" containerID="da060674cc453e1bf230796ab7551c2f212f5aa5acbe6b3db999bcad49461d13" exitCode=2 Dec 10 14:26:44 crc kubenswrapper[4847]: I1210 14:26:44.408825 4847 scope.go:117] "RemoveContainer" containerID="da060674cc453e1bf230796ab7551c2f212f5aa5acbe6b3db999bcad49461d13" Dec 10 14:26:45 crc kubenswrapper[4847]: I1210 14:26:45.619221 4847 patch_prober.go:28] interesting pod/console-f9d7485db-m2sr6 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 10 14:26:45 crc kubenswrapper[4847]: I1210 14:26:45.619608 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2sr6" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 10 14:26:47 crc kubenswrapper[4847]: I1210 14:26:47.160306 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-c6lv4" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.232538 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.353195 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir\") pod \"98581445-9906-4999-8a55-0e7e6fb218b6\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.353284 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access\") pod \"98581445-9906-4999-8a55-0e7e6fb218b6\" (UID: \"98581445-9906-4999-8a55-0e7e6fb218b6\") " Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.354198 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "98581445-9906-4999-8a55-0e7e6fb218b6" (UID: "98581445-9906-4999-8a55-0e7e6fb218b6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.360280 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "98581445-9906-4999-8a55-0e7e6fb218b6" (UID: "98581445-9906-4999-8a55-0e7e6fb218b6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.454917 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98581445-9906-4999-8a55-0e7e6fb218b6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.455206 4847 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/98581445-9906-4999-8a55-0e7e6fb218b6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.459113 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"98581445-9906-4999-8a55-0e7e6fb218b6","Type":"ContainerDied","Data":"6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4"} Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.459155 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a8d181c78a15e4fbbf2ce9b2695e128b02c2f1538781ae2e970ebfcff20b8c4" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.459237 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.464855 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-pcnwq_d21a2da8-a31f-40a3-84b5-db9d331e822e/cluster-samples-operator/0.log" Dec 10 14:26:52 crc kubenswrapper[4847]: I1210 14:26:52.464930 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pcnwq" event={"ID":"d21a2da8-a31f-40a3-84b5-db9d331e822e","Type":"ContainerStarted","Data":"e07c5fc8388efe60a2aba656a09608fed822cea0fdc49f0143ba080e99215eba"} Dec 10 14:26:53 crc kubenswrapper[4847]: I1210 14:26:53.511103 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:26:53 crc kubenswrapper[4847]: I1210 14:26:53.571888 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:53 crc kubenswrapper[4847]: I1210 14:26:53.577390 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e4db231f-2f4a-40e0-8137-c45b3f604fb6-metrics-certs\") pod \"network-metrics-daemon-2khbq\" (UID: \"e4db231f-2f4a-40e0-8137-c45b3f604fb6\") " pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:53 crc kubenswrapper[4847]: I1210 14:26:53.871920 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2khbq" Dec 10 14:26:55 crc kubenswrapper[4847]: I1210 14:26:55.627375 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:26:55 crc kubenswrapper[4847]: I1210 14:26:55.631217 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:27:01 crc kubenswrapper[4847]: I1210 14:27:01.010892 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:27:01 crc kubenswrapper[4847]: I1210 14:27:01.011012 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:27:07 crc kubenswrapper[4847]: I1210 14:27:07.630138 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfqtv" Dec 10 14:27:08 crc kubenswrapper[4847]: E1210 14:27:08.382789 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 14:27:08 crc kubenswrapper[4847]: E1210 14:27:08.382980 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wj6vw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kvfs8_openshift-marketplace(a6f20b79-6bb4-4832-a607-88805371713c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:08 crc kubenswrapper[4847]: E1210 14:27:08.384145 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kvfs8" podUID="a6f20b79-6bb4-4832-a607-88805371713c" Dec 10 14:27:09 crc kubenswrapper[4847]: E1210 14:27:09.544351 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kvfs8" podUID="a6f20b79-6bb4-4832-a607-88805371713c" Dec 10 14:27:10 crc kubenswrapper[4847]: E1210 14:27:10.263636 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 14:27:10 crc kubenswrapper[4847]: E1210 14:27:10.263849 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f9j64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4tsg4_openshift-marketplace(ddfe3352-42db-46fd-9e40-b4f4bf5f3292): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:10 crc kubenswrapper[4847]: E1210 14:27:10.265093 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4tsg4" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.195523 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 14:27:12 crc kubenswrapper[4847]: E1210 14:27:12.196157 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98581445-9906-4999-8a55-0e7e6fb218b6" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.196172 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="98581445-9906-4999-8a55-0e7e6fb218b6" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: E1210 14:27:12.196185 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="118f4abc-3653-4852-8f11-074074102527" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.196192 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="118f4abc-3653-4852-8f11-074074102527" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.196312 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="118f4abc-3653-4852-8f11-074074102527" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.196328 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="98581445-9906-4999-8a55-0e7e6fb218b6" containerName="pruner" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.196755 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.208035 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.208246 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.216268 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.327247 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.327278 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.428547 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.428646 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.428738 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.448302 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:12 crc kubenswrapper[4847]: I1210 14:27:12.523548 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:15 crc kubenswrapper[4847]: E1210 14:27:15.562060 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4tsg4" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" Dec 10 14:27:16 crc kubenswrapper[4847]: I1210 14:27:16.874838 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.481925 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.482103 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6sq9b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-8gv4z_openshift-marketplace(d73279bb-7be5-4397-9aee-ac26a8861a58): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.483541 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-8gv4z" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.562700 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.562882 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrk4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-kvx4r_openshift-marketplace(f681c8b8-c1fb-47a9-91c2-9122d0799eb1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.565112 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-kvx4r" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.665342 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.665492 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g7mhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-d69ss_openshift-marketplace(52e58c9f-f9db-48ac-9055-f32856b4b946): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.666676 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-d69ss" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.771093 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.771272 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5g4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kqrfk_openshift-marketplace(1629f237-abc0-4374-822b-d5b0cd0c9d62): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:17 crc kubenswrapper[4847]: E1210 14:27:17.772776 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kqrfk" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" Dec 10 14:27:17 crc kubenswrapper[4847]: I1210 14:27:17.994690 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 14:27:17 crc kubenswrapper[4847]: I1210 14:27:17.995552 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.002516 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.114453 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.114502 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.114569 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.215943 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.216003 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.216037 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.216162 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.216564 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.234729 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access\") pod \"installer-9-crc\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.260887 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-8gv4z" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.260915 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-kvx4r" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.327075 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.327245 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ksdhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-sm78p_openshift-marketplace(3987f250-b286-4eed-86e0-2c9e6abd9e3f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.328429 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-sm78p" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.350276 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.350425 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sxtnv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-q6kb7_openshift-marketplace(c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.355265 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-q6kb7" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.361025 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.520685 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2khbq"] Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.585562 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 14:27:18 crc kubenswrapper[4847]: W1210 14:27:18.592346 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda782016b_f55c_4dad_9617_19a0d4f0f42e.slice/crio-f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40 WatchSource:0}: Error finding container f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40: Status 404 returned error can't find the container with id f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40 Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.623312 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a782016b-f55c-4dad-9617-19a0d4f0f42e","Type":"ContainerStarted","Data":"f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40"} Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.624323 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2khbq" event={"ID":"e4db231f-2f4a-40e0-8137-c45b3f604fb6","Type":"ContainerStarted","Data":"669fbbd5d8971a6c0cb1da3f096f7ca5deec54e0bad69f2108fb23431ce89512"} Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.625526 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-sm78p" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.625983 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-q6kb7" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.626048 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-d69ss" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" Dec 10 14:27:18 crc kubenswrapper[4847]: E1210 14:27:18.631059 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kqrfk" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" Dec 10 14:27:18 crc kubenswrapper[4847]: I1210 14:27:18.680221 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 14:27:18 crc kubenswrapper[4847]: W1210 14:27:18.685013 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode9b1255d_044f_496a_b79c_72469d3d408e.slice/crio-d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7 WatchSource:0}: Error finding container d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7: Status 404 returned error can't find the container with id d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7 Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.631467 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2khbq" event={"ID":"e4db231f-2f4a-40e0-8137-c45b3f604fb6","Type":"ContainerStarted","Data":"7e5eef4ae9e4b42171528c60649d4feef86f6646994caf026092a948c2ee8df2"} Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.631778 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2khbq" event={"ID":"e4db231f-2f4a-40e0-8137-c45b3f604fb6","Type":"ContainerStarted","Data":"5d3c952810c85b37590350b55c86434cb41d5230745ed1440c1dd52656b9d24a"} Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.634817 4847 generic.go:334] "Generic (PLEG): container finished" podID="e9b1255d-044f-496a-b79c-72469d3d408e" containerID="69997f1ad7ece8ce5c03eb1c28b281fcc9e5e7e9b7ccb216ef3e4ca9360064d4" exitCode=0 Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.635040 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e9b1255d-044f-496a-b79c-72469d3d408e","Type":"ContainerDied","Data":"69997f1ad7ece8ce5c03eb1c28b281fcc9e5e7e9b7ccb216ef3e4ca9360064d4"} Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.635089 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e9b1255d-044f-496a-b79c-72469d3d408e","Type":"ContainerStarted","Data":"d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7"} Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.637280 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a782016b-f55c-4dad-9617-19a0d4f0f42e","Type":"ContainerStarted","Data":"830948ae04d3b3da9f4929f3ab1ad8c5cc5bdb8b525a3721d0c20150065b2236"} Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.646504 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2khbq" podStartSLOduration=168.646485758 podStartE2EDuration="2m48.646485758s" podCreationTimestamp="2025-12-10 14:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:27:19.644693257 +0000 UTC m=+189.213910907" watchObservedRunningTime="2025-12-10 14:27:19.646485758 +0000 UTC m=+189.215703388" Dec 10 14:27:19 crc kubenswrapper[4847]: I1210 14:27:19.679820 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.67980108 podStartE2EDuration="2.67980108s" podCreationTimestamp="2025-12-10 14:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:27:19.667618078 +0000 UTC m=+189.236835738" watchObservedRunningTime="2025-12-10 14:27:19.67980108 +0000 UTC m=+189.249018710" Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.853934 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.957755 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir\") pod \"e9b1255d-044f-496a-b79c-72469d3d408e\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.957864 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e9b1255d-044f-496a-b79c-72469d3d408e" (UID: "e9b1255d-044f-496a-b79c-72469d3d408e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.957906 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access\") pod \"e9b1255d-044f-496a-b79c-72469d3d408e\" (UID: \"e9b1255d-044f-496a-b79c-72469d3d408e\") " Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.958246 4847 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9b1255d-044f-496a-b79c-72469d3d408e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:20 crc kubenswrapper[4847]: I1210 14:27:20.963450 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e9b1255d-044f-496a-b79c-72469d3d408e" (UID: "e9b1255d-044f-496a-b79c-72469d3d408e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:21 crc kubenswrapper[4847]: I1210 14:27:21.059212 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9b1255d-044f-496a-b79c-72469d3d408e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:21 crc kubenswrapper[4847]: I1210 14:27:21.650162 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e9b1255d-044f-496a-b79c-72469d3d408e","Type":"ContainerDied","Data":"d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7"} Dec 10 14:27:21 crc kubenswrapper[4847]: I1210 14:27:21.650215 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d847a7c19f091a949f07f222921590c4c60237ce7b0d67a8f63c8abb405a4bb7" Dec 10 14:27:21 crc kubenswrapper[4847]: I1210 14:27:21.650517 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 14:27:24 crc kubenswrapper[4847]: I1210 14:27:24.670965 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6f20b79-6bb4-4832-a607-88805371713c" containerID="b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861" exitCode=0 Dec 10 14:27:24 crc kubenswrapper[4847]: I1210 14:27:24.671577 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerDied","Data":"b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861"} Dec 10 14:27:25 crc kubenswrapper[4847]: I1210 14:27:25.679218 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerStarted","Data":"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29"} Dec 10 14:27:25 crc kubenswrapper[4847]: I1210 14:27:25.699886 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kvfs8" podStartSLOduration=2.526585953 podStartE2EDuration="52.699868795s" podCreationTimestamp="2025-12-10 14:26:33 +0000 UTC" firstStartedPulling="2025-12-10 14:26:35.003150442 +0000 UTC m=+144.572368072" lastFinishedPulling="2025-12-10 14:27:25.176433284 +0000 UTC m=+194.745650914" observedRunningTime="2025-12-10 14:27:25.697620342 +0000 UTC m=+195.266837982" watchObservedRunningTime="2025-12-10 14:27:25.699868795 +0000 UTC m=+195.269086425" Dec 10 14:27:29 crc kubenswrapper[4847]: I1210 14:27:29.699269 4847 generic.go:334] "Generic (PLEG): container finished" podID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerID="11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019" exitCode=0 Dec 10 14:27:29 crc kubenswrapper[4847]: I1210 14:27:29.699346 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerDied","Data":"11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019"} Dec 10 14:27:30 crc kubenswrapper[4847]: I1210 14:27:30.712277 4847 generic.go:334] "Generic (PLEG): container finished" podID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerID="b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa" exitCode=0 Dec 10 14:27:30 crc kubenswrapper[4847]: I1210 14:27:30.712360 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerDied","Data":"b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa"} Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.011136 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.011435 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.719349 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerStarted","Data":"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906"} Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.721918 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerStarted","Data":"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587"} Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.724019 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerStarted","Data":"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f"} Dec 10 14:27:31 crc kubenswrapper[4847]: I1210 14:27:31.743667 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4tsg4" podStartSLOduration=3.400665919 podStartE2EDuration="56.743647426s" podCreationTimestamp="2025-12-10 14:26:35 +0000 UTC" firstStartedPulling="2025-12-10 14:26:38.139359379 +0000 UTC m=+147.708577009" lastFinishedPulling="2025-12-10 14:27:31.482340886 +0000 UTC m=+201.051558516" observedRunningTime="2025-12-10 14:27:31.740947186 +0000 UTC m=+201.310164816" watchObservedRunningTime="2025-12-10 14:27:31.743647426 +0000 UTC m=+201.312865056" Dec 10 14:27:32 crc kubenswrapper[4847]: I1210 14:27:32.731070 4847 generic.go:334] "Generic (PLEG): container finished" podID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerID="3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587" exitCode=0 Dec 10 14:27:32 crc kubenswrapper[4847]: I1210 14:27:32.731188 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerDied","Data":"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587"} Dec 10 14:27:32 crc kubenswrapper[4847]: I1210 14:27:32.733836 4847 generic.go:334] "Generic (PLEG): container finished" podID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerID="946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f" exitCode=0 Dec 10 14:27:32 crc kubenswrapper[4847]: I1210 14:27:32.734061 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerDied","Data":"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f"} Dec 10 14:27:32 crc kubenswrapper[4847]: I1210 14:27:32.735679 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerStarted","Data":"3ea1317a5fabbb133a153fe32cdb11359ed4f71796a3d362b3472ac6cc7b6136"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.742814 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerStarted","Data":"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.744938 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerStarted","Data":"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.746479 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerStarted","Data":"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.748483 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerStarted","Data":"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.750005 4847 generic.go:334] "Generic (PLEG): container finished" podID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerID="3ea1317a5fabbb133a153fe32cdb11359ed4f71796a3d362b3472ac6cc7b6136" exitCode=0 Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.750039 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerDied","Data":"3ea1317a5fabbb133a153fe32cdb11359ed4f71796a3d362b3472ac6cc7b6136"} Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.761460 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8gv4z" podStartSLOduration=2.725879941 podStartE2EDuration="57.761438039s" podCreationTimestamp="2025-12-10 14:26:36 +0000 UTC" firstStartedPulling="2025-12-10 14:26:38.14686533 +0000 UTC m=+147.716082970" lastFinishedPulling="2025-12-10 14:27:33.182423438 +0000 UTC m=+202.751641068" observedRunningTime="2025-12-10 14:27:33.759906114 +0000 UTC m=+203.329123754" watchObservedRunningTime="2025-12-10 14:27:33.761438039 +0000 UTC m=+203.330655669" Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.835702 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kvx4r" podStartSLOduration=2.133484124 podStartE2EDuration="59.835682129s" podCreationTimestamp="2025-12-10 14:26:34 +0000 UTC" firstStartedPulling="2025-12-10 14:26:35.016869836 +0000 UTC m=+144.586087466" lastFinishedPulling="2025-12-10 14:27:32.719067841 +0000 UTC m=+202.288285471" observedRunningTime="2025-12-10 14:27:33.830845576 +0000 UTC m=+203.400063216" watchObservedRunningTime="2025-12-10 14:27:33.835682129 +0000 UTC m=+203.404899759" Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.875537 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q6kb7" podStartSLOduration=1.758459954 podStartE2EDuration="59.875510219s" podCreationTimestamp="2025-12-10 14:26:34 +0000 UTC" firstStartedPulling="2025-12-10 14:26:35.034979772 +0000 UTC m=+144.604197402" lastFinishedPulling="2025-12-10 14:27:33.152030037 +0000 UTC m=+202.721247667" observedRunningTime="2025-12-10 14:27:33.871221442 +0000 UTC m=+203.440439092" watchObservedRunningTime="2025-12-10 14:27:33.875510219 +0000 UTC m=+203.444727849" Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.959035 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:27:33 crc kubenswrapper[4847]: I1210 14:27:33.959170 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.022629 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.341409 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.341932 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.566610 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.566839 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.766509 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerStarted","Data":"099b832b524a7ab36e3b25e9ac9db0cbf4df3056ba45993000bd5a47e21fe3f5"} Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.767419 4847 generic.go:334] "Generic (PLEG): container finished" podID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerID="f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1" exitCode=0 Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.767511 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerDied","Data":"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1"} Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.773525 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerStarted","Data":"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480"} Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.790507 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kqrfk" podStartSLOduration=2.825887759 podStartE2EDuration="57.790490804s" podCreationTimestamp="2025-12-10 14:26:37 +0000 UTC" firstStartedPulling="2025-12-10 14:26:39.219270948 +0000 UTC m=+148.788488578" lastFinishedPulling="2025-12-10 14:27:34.183873993 +0000 UTC m=+203.753091623" observedRunningTime="2025-12-10 14:27:34.789924077 +0000 UTC m=+204.359141717" watchObservedRunningTime="2025-12-10 14:27:34.790490804 +0000 UTC m=+204.359708434" Dec 10 14:27:34 crc kubenswrapper[4847]: I1210 14:27:34.811969 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:27:35 crc kubenswrapper[4847]: I1210 14:27:35.377012 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kvx4r" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="registry-server" probeResult="failure" output=< Dec 10 14:27:35 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:27:35 crc kubenswrapper[4847]: > Dec 10 14:27:35 crc kubenswrapper[4847]: I1210 14:27:35.612112 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-q6kb7" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="registry-server" probeResult="failure" output=< Dec 10 14:27:35 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:27:35 crc kubenswrapper[4847]: > Dec 10 14:27:35 crc kubenswrapper[4847]: I1210 14:27:35.780199 4847 generic.go:334] "Generic (PLEG): container finished" podID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerID="4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480" exitCode=0 Dec 10 14:27:35 crc kubenswrapper[4847]: I1210 14:27:35.780275 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerDied","Data":"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480"} Dec 10 14:27:36 crc kubenswrapper[4847]: I1210 14:27:36.189778 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:27:36 crc kubenswrapper[4847]: I1210 14:27:36.190122 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:27:36 crc kubenswrapper[4847]: I1210 14:27:36.230231 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:27:36 crc kubenswrapper[4847]: I1210 14:27:36.833691 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:27:37 crc kubenswrapper[4847]: I1210 14:27:37.185807 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:27:37 crc kubenswrapper[4847]: I1210 14:27:37.185853 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:27:37 crc kubenswrapper[4847]: I1210 14:27:37.629739 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:37 crc kubenswrapper[4847]: I1210 14:27:37.629787 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:37 crc kubenswrapper[4847]: I1210 14:27:37.792048 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerStarted","Data":"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1"} Dec 10 14:27:38 crc kubenswrapper[4847]: I1210 14:27:38.226645 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8gv4z" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="registry-server" probeResult="failure" output=< Dec 10 14:27:38 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:27:38 crc kubenswrapper[4847]: > Dec 10 14:27:38 crc kubenswrapper[4847]: I1210 14:27:38.679072 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kqrfk" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="registry-server" probeResult="failure" output=< Dec 10 14:27:38 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:27:38 crc kubenswrapper[4847]: > Dec 10 14:27:38 crc kubenswrapper[4847]: I1210 14:27:38.799235 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerStarted","Data":"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc"} Dec 10 14:27:38 crc kubenswrapper[4847]: I1210 14:27:38.816982 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d69ss" podStartSLOduration=2.97572151 podStartE2EDuration="1m2.816950599s" podCreationTimestamp="2025-12-10 14:26:36 +0000 UTC" firstStartedPulling="2025-12-10 14:26:38.166391696 +0000 UTC m=+147.735609326" lastFinishedPulling="2025-12-10 14:27:38.007620785 +0000 UTC m=+207.576838415" observedRunningTime="2025-12-10 14:27:38.815065493 +0000 UTC m=+208.384283123" watchObservedRunningTime="2025-12-10 14:27:38.816950599 +0000 UTC m=+208.386168229" Dec 10 14:27:38 crc kubenswrapper[4847]: I1210 14:27:38.819881 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sm78p" podStartSLOduration=4.311466865 podStartE2EDuration="1m5.819870876s" podCreationTimestamp="2025-12-10 14:26:33 +0000 UTC" firstStartedPulling="2025-12-10 14:26:35.031181056 +0000 UTC m=+144.600398686" lastFinishedPulling="2025-12-10 14:27:36.539585067 +0000 UTC m=+206.108802697" observedRunningTime="2025-12-10 14:27:37.813445423 +0000 UTC m=+207.382663053" watchObservedRunningTime="2025-12-10 14:27:38.819870876 +0000 UTC m=+208.389088536" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.155855 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.156874 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.203702 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.384958 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.442274 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.617825 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.662911 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:44 crc kubenswrapper[4847]: I1210 14:27:44.875791 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:27:45 crc kubenswrapper[4847]: I1210 14:27:45.131587 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:27:45 crc kubenswrapper[4847]: I1210 14:27:45.840364 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kvx4r" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="registry-server" containerID="cri-o://79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86" gracePeriod=2 Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.619806 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.619857 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.660154 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.881390 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.926446 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:27:46 crc kubenswrapper[4847]: I1210 14:27:46.926730 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q6kb7" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="registry-server" containerID="cri-o://d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c" gracePeriod=2 Dec 10 14:27:47 crc kubenswrapper[4847]: I1210 14:27:47.238805 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:27:47 crc kubenswrapper[4847]: I1210 14:27:47.290550 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:27:47 crc kubenswrapper[4847]: I1210 14:27:47.665758 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:47 crc kubenswrapper[4847]: I1210 14:27:47.702799 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:49 crc kubenswrapper[4847]: I1210 14:27:49.330553 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:27:49 crc kubenswrapper[4847]: I1210 14:27:49.331017 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d69ss" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="registry-server" containerID="cri-o://aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc" gracePeriod=2 Dec 10 14:27:49 crc kubenswrapper[4847]: I1210 14:27:49.854692 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht7zw"] Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.261675 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.266732 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q6kb7_c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0/registry-server/0.log" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.273886 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.349342 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content\") pod \"52e58c9f-f9db-48ac-9055-f32856b4b946\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.350024 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxtnv\" (UniqueName: \"kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv\") pod \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.350131 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7mhw\" (UniqueName: \"kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw\") pod \"52e58c9f-f9db-48ac-9055-f32856b4b946\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.350188 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities\") pod \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.350219 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content\") pod \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\" (UID: \"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.350248 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities\") pod \"52e58c9f-f9db-48ac-9055-f32856b4b946\" (UID: \"52e58c9f-f9db-48ac-9055-f32856b4b946\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.351623 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities" (OuterVolumeSpecName: "utilities") pod "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" (UID: "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.351658 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities" (OuterVolumeSpecName: "utilities") pod "52e58c9f-f9db-48ac-9055-f32856b4b946" (UID: "52e58c9f-f9db-48ac-9055-f32856b4b946"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.356721 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw" (OuterVolumeSpecName: "kube-api-access-g7mhw") pod "52e58c9f-f9db-48ac-9055-f32856b4b946" (UID: "52e58c9f-f9db-48ac-9055-f32856b4b946"). InnerVolumeSpecName "kube-api-access-g7mhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.356807 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv" (OuterVolumeSpecName: "kube-api-access-sxtnv") pod "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" (UID: "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0"). InnerVolumeSpecName "kube-api-access-sxtnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.384478 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52e58c9f-f9db-48ac-9055-f32856b4b946" (UID: "52e58c9f-f9db-48ac-9055-f32856b4b946"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.411899 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" (UID: "c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.451977 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxtnv\" (UniqueName: \"kubernetes.io/projected/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-kube-api-access-sxtnv\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.452017 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7mhw\" (UniqueName: \"kubernetes.io/projected/52e58c9f-f9db-48ac-9055-f32856b4b946-kube-api-access-g7mhw\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.452066 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.452079 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.452092 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.452103 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e58c9f-f9db-48ac-9055-f32856b4b946-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.760227 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kvx4r_f681c8b8-c1fb-47a9-91c2-9122d0799eb1/registry-server/0.log" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.763543 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.855933 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities\") pod \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.856106 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content\") pod \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.856189 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrk4j\" (UniqueName: \"kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j\") pod \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\" (UID: \"f681c8b8-c1fb-47a9-91c2-9122d0799eb1\") " Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.857062 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities" (OuterVolumeSpecName: "utilities") pod "f681c8b8-c1fb-47a9-91c2-9122d0799eb1" (UID: "f681c8b8-c1fb-47a9-91c2-9122d0799eb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.859815 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j" (OuterVolumeSpecName: "kube-api-access-wrk4j") pod "f681c8b8-c1fb-47a9-91c2-9122d0799eb1" (UID: "f681c8b8-c1fb-47a9-91c2-9122d0799eb1"). InnerVolumeSpecName "kube-api-access-wrk4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.872897 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kvx4r_f681c8b8-c1fb-47a9-91c2-9122d0799eb1/registry-server/0.log" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.873693 4847 generic.go:334] "Generic (PLEG): container finished" podID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerID="79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86" exitCode=137 Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.873768 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerDied","Data":"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.873800 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvx4r" event={"ID":"f681c8b8-c1fb-47a9-91c2-9122d0799eb1","Type":"ContainerDied","Data":"21e448667cfbb3aaab4ba11e9f3f132784f1fab6e0f6d144beeac7169c5244c3"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.873818 4847 scope.go:117] "RemoveContainer" containerID="79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.873941 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvx4r" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.877321 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-q6kb7_c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0/registry-server/0.log" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.877948 4847 generic.go:334] "Generic (PLEG): container finished" podID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerID="d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c" exitCode=137 Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.878000 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerDied","Data":"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.878024 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q6kb7" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.878042 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q6kb7" event={"ID":"c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0","Type":"ContainerDied","Data":"e3c6e07680c90b57ca6913b00d94eded585449c3e8cae10eae1036d015002ddb"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.880123 4847 generic.go:334] "Generic (PLEG): container finished" podID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerID="aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc" exitCode=0 Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.880159 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d69ss" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.880161 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerDied","Data":"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.880437 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d69ss" event={"ID":"52e58c9f-f9db-48ac-9055-f32856b4b946","Type":"ContainerDied","Data":"6123b75dfae074abd332cfbc6e9e0ed3e31c71f6eb93f00da4810af44e083eab"} Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.899201 4847 scope.go:117] "RemoveContainer" containerID="b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.904552 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.913767 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q6kb7"] Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.917087 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.917975 4847 scope.go:117] "RemoveContainer" containerID="ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.919548 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d69ss"] Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.923348 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f681c8b8-c1fb-47a9-91c2-9122d0799eb1" (UID: "f681c8b8-c1fb-47a9-91c2-9122d0799eb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.933081 4847 scope.go:117] "RemoveContainer" containerID="79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.933628 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86\": container with ID starting with 79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86 not found: ID does not exist" containerID="79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.933702 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86"} err="failed to get container status \"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86\": rpc error: code = NotFound desc = could not find container \"79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86\": container with ID starting with 79721885cd3cac7cbf14458fd965d7896b71f02442926230e81b9766208dec86 not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.933821 4847 scope.go:117] "RemoveContainer" containerID="b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.934222 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa\": container with ID starting with b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa not found: ID does not exist" containerID="b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.934281 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa"} err="failed to get container status \"b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa\": rpc error: code = NotFound desc = could not find container \"b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa\": container with ID starting with b10d861ea8a31e86b4a71f9c67156582a93f1b67f135ccb2cf69b2483bdffffa not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.934303 4847 scope.go:117] "RemoveContainer" containerID="ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.934880 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191\": container with ID starting with ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191 not found: ID does not exist" containerID="ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.934939 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191"} err="failed to get container status \"ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191\": rpc error: code = NotFound desc = could not find container \"ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191\": container with ID starting with ee3d48e75789d88cfdbb72b681d033d0d20952e99a58fbe7d05cbba8e3f8a191 not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.934973 4847 scope.go:117] "RemoveContainer" containerID="d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.948786 4847 scope.go:117] "RemoveContainer" containerID="946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.957330 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.957357 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.957369 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrk4j\" (UniqueName: \"kubernetes.io/projected/f681c8b8-c1fb-47a9-91c2-9122d0799eb1-kube-api-access-wrk4j\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.966501 4847 scope.go:117] "RemoveContainer" containerID="0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.980743 4847 scope.go:117] "RemoveContainer" containerID="d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.981113 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c\": container with ID starting with d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c not found: ID does not exist" containerID="d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.981166 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c"} err="failed to get container status \"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c\": rpc error: code = NotFound desc = could not find container \"d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c\": container with ID starting with d019abc3ffc394dd9f2caecc0da43058b7a7e914eb8f68aa94c14c9d8002318c not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.981200 4847 scope.go:117] "RemoveContainer" containerID="946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.981490 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f\": container with ID starting with 946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f not found: ID does not exist" containerID="946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.981522 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f"} err="failed to get container status \"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f\": rpc error: code = NotFound desc = could not find container \"946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f\": container with ID starting with 946c329f7f95b4ae023db33c6c6ad13147c691cb4b86f741a810031f1dc62b1f not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.981542 4847 scope.go:117] "RemoveContainer" containerID="0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95" Dec 10 14:27:50 crc kubenswrapper[4847]: E1210 14:27:50.981999 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95\": container with ID starting with 0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95 not found: ID does not exist" containerID="0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.982041 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95"} err="failed to get container status \"0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95\": rpc error: code = NotFound desc = could not find container \"0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95\": container with ID starting with 0906707b9471b7e6f7b3cfa9a77cb673cb19633850c85feda87d098adbaf4f95 not found: ID does not exist" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.982069 4847 scope.go:117] "RemoveContainer" containerID="aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc" Dec 10 14:27:50 crc kubenswrapper[4847]: I1210 14:27:50.993182 4847 scope.go:117] "RemoveContainer" containerID="4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.006622 4847 scope.go:117] "RemoveContainer" containerID="dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.018449 4847 scope.go:117] "RemoveContainer" containerID="aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc" Dec 10 14:27:51 crc kubenswrapper[4847]: E1210 14:27:51.018841 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc\": container with ID starting with aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc not found: ID does not exist" containerID="aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.018893 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc"} err="failed to get container status \"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc\": rpc error: code = NotFound desc = could not find container \"aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc\": container with ID starting with aaa0f823dd8cd3400d671244852c253cef05aa450172ac86c475a78138f9debc not found: ID does not exist" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.018914 4847 scope.go:117] "RemoveContainer" containerID="4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480" Dec 10 14:27:51 crc kubenswrapper[4847]: E1210 14:27:51.019311 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480\": container with ID starting with 4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480 not found: ID does not exist" containerID="4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.019335 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480"} err="failed to get container status \"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480\": rpc error: code = NotFound desc = could not find container \"4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480\": container with ID starting with 4697f5f9f782f90d40e017a4e90f58eedac37a35626651ac2d7e37c55840a480 not found: ID does not exist" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.019349 4847 scope.go:117] "RemoveContainer" containerID="dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc" Dec 10 14:27:51 crc kubenswrapper[4847]: E1210 14:27:51.019625 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc\": container with ID starting with dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc not found: ID does not exist" containerID="dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.019669 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc"} err="failed to get container status \"dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc\": rpc error: code = NotFound desc = could not find container \"dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc\": container with ID starting with dccd848cff7387ba19b203c74e63200c0ff7c7c73bafea039ecd1c0f9c283adc not found: ID does not exist" Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.203097 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.205424 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kvx4r"] Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.729809 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.730420 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kqrfk" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="registry-server" containerID="cri-o://099b832b524a7ab36e3b25e9ac9db0cbf4df3056ba45993000bd5a47e21fe3f5" gracePeriod=2 Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.902161 4847 generic.go:334] "Generic (PLEG): container finished" podID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerID="099b832b524a7ab36e3b25e9ac9db0cbf4df3056ba45993000bd5a47e21fe3f5" exitCode=0 Dec 10 14:27:51 crc kubenswrapper[4847]: I1210 14:27:51.902232 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerDied","Data":"099b832b524a7ab36e3b25e9ac9db0cbf4df3056ba45993000bd5a47e21fe3f5"} Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.111398 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.272729 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities\") pod \"1629f237-abc0-4374-822b-d5b0cd0c9d62\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.272827 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content\") pod \"1629f237-abc0-4374-822b-d5b0cd0c9d62\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.272884 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5g4l\" (UniqueName: \"kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l\") pod \"1629f237-abc0-4374-822b-d5b0cd0c9d62\" (UID: \"1629f237-abc0-4374-822b-d5b0cd0c9d62\") " Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.273549 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities" (OuterVolumeSpecName: "utilities") pod "1629f237-abc0-4374-822b-d5b0cd0c9d62" (UID: "1629f237-abc0-4374-822b-d5b0cd0c9d62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.276429 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l" (OuterVolumeSpecName: "kube-api-access-r5g4l") pod "1629f237-abc0-4374-822b-d5b0cd0c9d62" (UID: "1629f237-abc0-4374-822b-d5b0cd0c9d62"). InnerVolumeSpecName "kube-api-access-r5g4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.374620 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5g4l\" (UniqueName: \"kubernetes.io/projected/1629f237-abc0-4374-822b-d5b0cd0c9d62-kube-api-access-r5g4l\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.374672 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.381738 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1629f237-abc0-4374-822b-d5b0cd0c9d62" (UID: "1629f237-abc0-4374-822b-d5b0cd0c9d62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.476305 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1629f237-abc0-4374-822b-d5b0cd0c9d62-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.765549 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" path="/var/lib/kubelet/pods/52e58c9f-f9db-48ac-9055-f32856b4b946/volumes" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.766332 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" path="/var/lib/kubelet/pods/c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0/volumes" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.766880 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" path="/var/lib/kubelet/pods/f681c8b8-c1fb-47a9-91c2-9122d0799eb1/volumes" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.909699 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqrfk" event={"ID":"1629f237-abc0-4374-822b-d5b0cd0c9d62","Type":"ContainerDied","Data":"df1cc3e7362c7cb0db5a66bbe25ac6f8f5faa2c71733c1cd827446180b75ecf5"} Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.909813 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqrfk" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.910130 4847 scope.go:117] "RemoveContainer" containerID="099b832b524a7ab36e3b25e9ac9db0cbf4df3056ba45993000bd5a47e21fe3f5" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.929921 4847 scope.go:117] "RemoveContainer" containerID="3ea1317a5fabbb133a153fe32cdb11359ed4f71796a3d362b3472ac6cc7b6136" Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.936177 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.941769 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kqrfk"] Dec 10 14:27:52 crc kubenswrapper[4847]: I1210 14:27:52.947670 4847 scope.go:117] "RemoveContainer" containerID="f9f3622f1ea9c66bea169cf8e7b965a9959fc8f7457c1e1592f14e662fb63f77" Dec 10 14:27:54 crc kubenswrapper[4847]: I1210 14:27:54.767151 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" path="/var/lib/kubelet/pods/1629f237-abc0-4374-822b-d5b0cd0c9d62/volumes" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.410678 4847 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411374 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411386 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411403 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411410 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411418 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411424 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411432 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411438 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411448 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411455 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411465 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411470 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411478 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411484 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411492 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411498 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411508 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411513 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="extract-utilities" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411519 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411526 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411533 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411538 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="extract-content" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411546 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b1255d-044f-496a-b79c-72469d3d408e" containerName="pruner" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411552 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b1255d-044f-496a-b79c-72469d3d408e" containerName="pruner" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.411559 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411565 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411667 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f681c8b8-c1fb-47a9-91c2-9122d0799eb1" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411676 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9b1255d-044f-496a-b79c-72469d3d408e" containerName="pruner" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411685 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e58c9f-f9db-48ac-9055-f32856b4b946" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411694 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="1629f237-abc0-4374-822b-d5b0cd0c9d62" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.411702 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c714c7b0-d2ab-4efe-a2e6-8eec5d4da9d0" containerName="registry-server" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412036 4847 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412185 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412269 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6" gracePeriod=15 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412301 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99" gracePeriod=15 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412361 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f" gracePeriod=15 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412397 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b" gracePeriod=15 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412380 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f" gracePeriod=15 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.412859 4847 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413047 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413068 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413083 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413092 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413107 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413116 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413125 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413131 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413140 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413147 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.413157 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413163 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413283 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413303 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413313 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413322 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.413332 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.454380 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523675 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523755 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523792 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523809 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523828 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523848 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523867 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.523883 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625548 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625641 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625668 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625684 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625697 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625781 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625823 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625838 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625872 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625897 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625951 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.625928 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.626066 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.626192 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.626307 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.626423 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.759633 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:27:56 crc kubenswrapper[4847]: E1210 14:27:56.787883 4847 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fe0ea8386770e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 14:27:56.786120462 +0000 UTC m=+226.355338092,LastTimestamp:2025-12-10 14:27:56.786120462 +0000 UTC m=+226.355338092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.939931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"27b28f8aad1ddd187c9143f2d217b738fea3d84fd39c70e6fee509ce74460cc3"} Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.944080 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.944861 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99" exitCode=0 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.944894 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f" exitCode=0 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.944907 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f" exitCode=0 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.944922 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b" exitCode=2 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.946152 4847 generic.go:334] "Generic (PLEG): container finished" podID="a782016b-f55c-4dad-9617-19a0d4f0f42e" containerID="830948ae04d3b3da9f4929f3ab1ad8c5cc5bdb8b525a3721d0c20150065b2236" exitCode=0 Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.946190 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a782016b-f55c-4dad-9617-19a0d4f0f42e","Type":"ContainerDied","Data":"830948ae04d3b3da9f4929f3ab1ad8c5cc5bdb8b525a3721d0c20150065b2236"} Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.946733 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:56 crc kubenswrapper[4847]: I1210 14:27:56.946905 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:57 crc kubenswrapper[4847]: I1210 14:27:57.952474 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb"} Dec 10 14:27:57 crc kubenswrapper[4847]: I1210 14:27:57.953197 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:57 crc kubenswrapper[4847]: I1210 14:27:57.953644 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.150781 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.151644 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.152027 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.258754 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access\") pod \"a782016b-f55c-4dad-9617-19a0d4f0f42e\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.258803 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir\") pod \"a782016b-f55c-4dad-9617-19a0d4f0f42e\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.258836 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock\") pod \"a782016b-f55c-4dad-9617-19a0d4f0f42e\" (UID: \"a782016b-f55c-4dad-9617-19a0d4f0f42e\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.258901 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a782016b-f55c-4dad-9617-19a0d4f0f42e" (UID: "a782016b-f55c-4dad-9617-19a0d4f0f42e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.259054 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock" (OuterVolumeSpecName: "var-lock") pod "a782016b-f55c-4dad-9617-19a0d4f0f42e" (UID: "a782016b-f55c-4dad-9617-19a0d4f0f42e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.259215 4847 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.259234 4847 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a782016b-f55c-4dad-9617-19a0d4f0f42e-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.313821 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a782016b-f55c-4dad-9617-19a0d4f0f42e" (UID: "a782016b-f55c-4dad-9617-19a0d4f0f42e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.360941 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a782016b-f55c-4dad-9617-19a0d4f0f42e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.771769 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.773417 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.774289 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.774550 4847 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.774826 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865373 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865426 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865450 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865516 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865529 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865609 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865764 4847 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865780 4847 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.865790 4847 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.959017 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.959499 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a782016b-f55c-4dad-9617-19a0d4f0f42e","Type":"ContainerDied","Data":"f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40"} Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.959521 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6dde907ad91bf85ff2cfd9e11ac64af19e0e1e1cac4fdd6ddc8a44e78d33a40" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962172 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962329 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962556 4847 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962800 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962919 4847 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6" exitCode=0 Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.962999 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.963005 4847 scope.go:117] "RemoveContainer" containerID="f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.975123 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.975463 4847 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.976041 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.978046 4847 scope.go:117] "RemoveContainer" containerID="503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f" Dec 10 14:27:58 crc kubenswrapper[4847]: I1210 14:27:58.992039 4847 scope.go:117] "RemoveContainer" containerID="b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.009040 4847 scope.go:117] "RemoveContainer" containerID="d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.024736 4847 scope.go:117] "RemoveContainer" containerID="dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.041605 4847 scope.go:117] "RemoveContainer" containerID="74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.059974 4847 scope.go:117] "RemoveContainer" containerID="f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.060600 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\": container with ID starting with f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99 not found: ID does not exist" containerID="f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.060647 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99"} err="failed to get container status \"f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\": rpc error: code = NotFound desc = could not find container \"f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99\": container with ID starting with f7350584602b48443bcc29a1536c8f7cc6719bc95442ca5ed3df9f0be0d58a99 not found: ID does not exist" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.060679 4847 scope.go:117] "RemoveContainer" containerID="503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.061296 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\": container with ID starting with 503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f not found: ID does not exist" containerID="503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.061320 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f"} err="failed to get container status \"503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\": rpc error: code = NotFound desc = could not find container \"503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f\": container with ID starting with 503842bcfb72d1411a33be56c372b64ed82b36f06e9fc714d16f22922902788f not found: ID does not exist" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.061336 4847 scope.go:117] "RemoveContainer" containerID="b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.061756 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\": container with ID starting with b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f not found: ID does not exist" containerID="b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.061779 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f"} err="failed to get container status \"b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\": rpc error: code = NotFound desc = could not find container \"b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f\": container with ID starting with b25d998192199d58b93b7f3d07e37822bc0c63861029102f3b94c90ec519892f not found: ID does not exist" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.061796 4847 scope.go:117] "RemoveContainer" containerID="d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.062159 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\": container with ID starting with d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b not found: ID does not exist" containerID="d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.062183 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b"} err="failed to get container status \"d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\": rpc error: code = NotFound desc = could not find container \"d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b\": container with ID starting with d9506ffe43f0a63e61a2081a56312ba8bc01d6e305f626c0707290b2f5ebd86b not found: ID does not exist" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.062200 4847 scope.go:117] "RemoveContainer" containerID="dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.062485 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\": container with ID starting with dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6 not found: ID does not exist" containerID="dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.062590 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6"} err="failed to get container status \"dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\": rpc error: code = NotFound desc = could not find container \"dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6\": container with ID starting with dcc698c2a3ab21f81156744d4512f899d59946b4d12449948a66bb0bd2841db6 not found: ID does not exist" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.062612 4847 scope.go:117] "RemoveContainer" containerID="74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011" Dec 10 14:27:59 crc kubenswrapper[4847]: E1210 14:27:59.062867 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\": container with ID starting with 74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011 not found: ID does not exist" containerID="74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011" Dec 10 14:27:59 crc kubenswrapper[4847]: I1210 14:27:59.063039 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011"} err="failed to get container status \"74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\": rpc error: code = NotFound desc = could not find container \"74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011\": container with ID starting with 74d23062a944b27140fd3d689d31671b67a44dc54c030b7c7590375664828011 not found: ID does not exist" Dec 10 14:28:00 crc kubenswrapper[4847]: I1210 14:28:00.762515 4847 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:00 crc kubenswrapper[4847]: I1210 14:28:00.764107 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:00 crc kubenswrapper[4847]: I1210 14:28:00.764363 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:00 crc kubenswrapper[4847]: I1210 14:28:00.765950 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.011514 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.011584 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.011643 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.012261 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.012326 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7" gracePeriod=600 Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.980898 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7" exitCode=0 Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.980979 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7"} Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.981194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a"} Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.982315 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.982794 4847 status_manager.go:851] "Failed to get status for pod" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gvdwq\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:01 crc kubenswrapper[4847]: I1210 14:28:01.983186 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:02 crc kubenswrapper[4847]: E1210 14:28:02.209492 4847 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187fe0ea8386770e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 14:27:56.786120462 +0000 UTC m=+226.355338092,LastTimestamp:2025-12-10 14:27:56.786120462 +0000 UTC m=+226.355338092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.857029 4847 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.858331 4847 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.859348 4847 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.859876 4847 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.860536 4847 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:05 crc kubenswrapper[4847]: I1210 14:28:05.860645 4847 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 10 14:28:05 crc kubenswrapper[4847]: E1210 14:28:05.861199 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 10 14:28:06 crc kubenswrapper[4847]: E1210 14:28:06.061877 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 10 14:28:06 crc kubenswrapper[4847]: E1210 14:28:06.462890 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 10 14:28:07 crc kubenswrapper[4847]: E1210 14:28:07.264204 4847 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.759508 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.760262 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.760577 4847 status_manager.go:851] "Failed to get status for pod" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gvdwq\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.760879 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.772324 4847 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.772355 4847 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:07 crc kubenswrapper[4847]: E1210 14:28:07.772622 4847 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:07 crc kubenswrapper[4847]: I1210 14:28:07.773076 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:07 crc kubenswrapper[4847]: W1210 14:28:07.793660 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-36cebe2b3b107d4abbe27f291d9cd70fc7ee9c7830695ddbf7e48c1646dd9a05 WatchSource:0}: Error finding container 36cebe2b3b107d4abbe27f291d9cd70fc7ee9c7830695ddbf7e48c1646dd9a05: Status 404 returned error can't find the container with id 36cebe2b3b107d4abbe27f291d9cd70fc7ee9c7830695ddbf7e48c1646dd9a05 Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.015954 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"34f470b5eb16ba089cb3680784adcc75c3acfe1051c894b06b057f07f4518022"} Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.015999 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"36cebe2b3b107d4abbe27f291d9cd70fc7ee9c7830695ddbf7e48c1646dd9a05"} Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.016243 4847 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.016259 4847 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.016542 4847 status_manager.go:851] "Failed to get status for pod" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gvdwq\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:08 crc kubenswrapper[4847]: E1210 14:28:08.016629 4847 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.016842 4847 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:08 crc kubenswrapper[4847]: I1210 14:28:08.017165 4847 status_manager.go:851] "Failed to get status for pod" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.034973 4847 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="34f470b5eb16ba089cb3680784adcc75c3acfe1051c894b06b057f07f4518022" exitCode=0 Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035170 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"34f470b5eb16ba089cb3680784adcc75c3acfe1051c894b06b057f07f4518022"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035325 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63b714df875235ebc472ae2e3c8e931ac255307212aff909011d39bfc5d291fc"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035340 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f4d956a2d71c502050e515e78f060ffa2b28628fdd263ec49782f64be8a9f397"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035350 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3830f62d49b2c184367dc20263e4174edac5ebdf1d77ba20890d8e4da3aa296b"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035360 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8dcf71c801fa048eac645a1534f349ce07f573609830f4c77600dfcad96755e1"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035367 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5cb2a9ff1c16b2b0ddf79db9dfaa54b0714c9d1b01b1d3728c6083f9788f7784"} Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035608 4847 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035619 4847 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:09 crc kubenswrapper[4847]: I1210 14:28:09.035634 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:11 crc kubenswrapper[4847]: I1210 14:28:11.047369 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 14:28:11 crc kubenswrapper[4847]: I1210 14:28:11.047707 4847 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492" exitCode=1 Dec 10 14:28:11 crc kubenswrapper[4847]: I1210 14:28:11.047762 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492"} Dec 10 14:28:11 crc kubenswrapper[4847]: I1210 14:28:11.048279 4847 scope.go:117] "RemoveContainer" containerID="1f0dec2e3617d3dfc15f452a669a5df0a40953b78cc1ae588b9bd9693b9cc492" Dec 10 14:28:12 crc kubenswrapper[4847]: I1210 14:28:12.056886 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 14:28:12 crc kubenswrapper[4847]: I1210 14:28:12.056952 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"441680d7a18af08cd2befc58a84d8dc7c523e36b60b6d4dff0c55a6fd1f9c681"} Dec 10 14:28:12 crc kubenswrapper[4847]: I1210 14:28:12.773950 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:12 crc kubenswrapper[4847]: I1210 14:28:12.774330 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:12 crc kubenswrapper[4847]: I1210 14:28:12.782486 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:13 crc kubenswrapper[4847]: I1210 14:28:13.536589 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:28:14 crc kubenswrapper[4847]: I1210 14:28:14.772400 4847 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:14 crc kubenswrapper[4847]: I1210 14:28:14.861468 4847 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="18b485b2-cef5-4841-ba84-fedca00231f7" Dec 10 14:28:14 crc kubenswrapper[4847]: I1210 14:28:14.885653 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" podUID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" containerName="oauth-openshift" containerID="cri-o://448d2329f9c2e99923cd8f25a367e6863c758ea2cec9128aab5b265a7fccb01c" gracePeriod=15 Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.084601 4847 generic.go:334] "Generic (PLEG): container finished" podID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" containerID="448d2329f9c2e99923cd8f25a367e6863c758ea2cec9128aab5b265a7fccb01c" exitCode=0 Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.084686 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" event={"ID":"2e6718b4-c798-4e9a-9dff-fd3bcd9566db","Type":"ContainerDied","Data":"448d2329f9c2e99923cd8f25a367e6863c758ea2cec9128aab5b265a7fccb01c"} Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.085159 4847 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.085173 4847 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7c80a69d-2927-4296-bb72-9baec8edb31f" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.089688 4847 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="18b485b2-cef5-4841-ba84-fedca00231f7" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.226583 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379230 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379274 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379323 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmnk9\" (UniqueName: \"kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379346 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379363 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379388 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379424 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379440 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379472 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379495 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379512 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379532 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379554 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.379573 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session\") pod \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\" (UID: \"2e6718b4-c798-4e9a-9dff-fd3bcd9566db\") " Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.380376 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.381355 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.381369 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.381594 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.381679 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.385499 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.385759 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.386025 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.386331 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.386345 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9" (OuterVolumeSpecName: "kube-api-access-nmnk9") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "kube-api-access-nmnk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.387212 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.387254 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.388150 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.389447 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2e6718b4-c798-4e9a-9dff-fd3bcd9566db" (UID: "2e6718b4-c798-4e9a-9dff-fd3bcd9566db"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.481994 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482035 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482053 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482066 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482076 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482086 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482099 4847 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482109 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmnk9\" (UniqueName: \"kubernetes.io/projected/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-kube-api-access-nmnk9\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482118 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482144 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482161 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482172 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482181 4847 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:15 crc kubenswrapper[4847]: I1210 14:28:15.482193 4847 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e6718b4-c798-4e9a-9dff-fd3bcd9566db-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:16 crc kubenswrapper[4847]: I1210 14:28:16.091833 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" event={"ID":"2e6718b4-c798-4e9a-9dff-fd3bcd9566db","Type":"ContainerDied","Data":"e62805c05c7a6758b32378a78bd5fcd56cb4b8e34248d8c7128ac6758a96ba0a"} Dec 10 14:28:16 crc kubenswrapper[4847]: I1210 14:28:16.091899 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht7zw" Dec 10 14:28:16 crc kubenswrapper[4847]: I1210 14:28:16.091925 4847 scope.go:117] "RemoveContainer" containerID="448d2329f9c2e99923cd8f25a367e6863c758ea2cec9128aab5b265a7fccb01c" Dec 10 14:28:18 crc kubenswrapper[4847]: I1210 14:28:18.390125 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:28:18 crc kubenswrapper[4847]: I1210 14:28:18.394541 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:28:21 crc kubenswrapper[4847]: I1210 14:28:21.970956 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 14:28:22 crc kubenswrapper[4847]: I1210 14:28:22.034557 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 14:28:22 crc kubenswrapper[4847]: I1210 14:28:22.300449 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 14:28:22 crc kubenswrapper[4847]: I1210 14:28:22.375996 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 14:28:22 crc kubenswrapper[4847]: I1210 14:28:22.894155 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 14:28:23 crc kubenswrapper[4847]: I1210 14:28:23.133152 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 14:28:23 crc kubenswrapper[4847]: I1210 14:28:23.540095 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 14:28:23 crc kubenswrapper[4847]: I1210 14:28:23.969132 4847 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 14:28:24 crc kubenswrapper[4847]: I1210 14:28:24.110168 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 14:28:24 crc kubenswrapper[4847]: I1210 14:28:24.582830 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 14:28:24 crc kubenswrapper[4847]: I1210 14:28:24.941908 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 14:28:25 crc kubenswrapper[4847]: I1210 14:28:25.362272 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 14:28:25 crc kubenswrapper[4847]: I1210 14:28:25.434335 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 14:28:25 crc kubenswrapper[4847]: I1210 14:28:25.883864 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 14:28:26 crc kubenswrapper[4847]: I1210 14:28:26.146977 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 14:28:26 crc kubenswrapper[4847]: I1210 14:28:26.451735 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 14:28:26 crc kubenswrapper[4847]: I1210 14:28:26.747596 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 14:28:26 crc kubenswrapper[4847]: I1210 14:28:26.949651 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 14:28:26 crc kubenswrapper[4847]: I1210 14:28:26.975594 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.242964 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.357357 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.413498 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.434489 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.830680 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 14:28:27 crc kubenswrapper[4847]: I1210 14:28:27.941997 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.089829 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.312451 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.487038 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.635403 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.866947 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 14:28:28 crc kubenswrapper[4847]: I1210 14:28:28.959922 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.189505 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.226218 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.297489 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.334028 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.368317 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.632399 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.808963 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.809799 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.832322 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.835256 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 14:28:29 crc kubenswrapper[4847]: I1210 14:28:29.907068 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.026873 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.035106 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.058555 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.246672 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.317468 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.434551 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.447019 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.450918 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.479419 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.492545 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.545335 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.584881 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.703967 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.769561 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.787608 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 14:28:30 crc kubenswrapper[4847]: I1210 14:28:30.821957 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.067286 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.117144 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.134623 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.181915 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.251467 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.737439 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.793024 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.871784 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.934121 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 14:28:31 crc kubenswrapper[4847]: I1210 14:28:31.975547 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.007097 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.121678 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.184579 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.248556 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.293548 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.294500 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.316630 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.403699 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.479356 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.479390 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.490168 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.603095 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.625539 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.666924 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.725505 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.936591 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.975239 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.976990 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 14:28:32 crc kubenswrapper[4847]: I1210 14:28:32.996425 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.007052 4847 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.071884 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.081358 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.130738 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.181869 4847 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.239642 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.248730 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.271694 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.317336 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.336618 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.394603 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.525101 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.658047 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.686964 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.714447 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.763805 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.802420 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.802420 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.808351 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.811258 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.875535 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 14:28:33 crc kubenswrapper[4847]: I1210 14:28:33.882526 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.036943 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.037786 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.039232 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.062399 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.097695 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.198106 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.232983 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.283252 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.294937 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.358160 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.371442 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.382421 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.606578 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.635410 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.689582 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.754021 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.847418 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.929286 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 14:28:34 crc kubenswrapper[4847]: I1210 14:28:34.948761 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.019287 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.061842 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.122941 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.159226 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.180656 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.270066 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.349343 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.439108 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.475931 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.566203 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.612126 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.688127 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.727623 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.796562 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.836688 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 14:28:35 crc kubenswrapper[4847]: I1210 14:28:35.889083 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.068329 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.077839 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.138651 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.215545 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.232029 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.303384 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.436386 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.447285 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.478462 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.539261 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.563841 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.581872 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.665194 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.677367 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.732322 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.784066 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.839054 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 14:28:36 crc kubenswrapper[4847]: I1210 14:28:36.852509 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.077876 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.368667 4847 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.372005 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.371990994 podStartE2EDuration="41.371990994s" podCreationTimestamp="2025-12-10 14:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:28:14.834640662 +0000 UTC m=+244.403858292" watchObservedRunningTime="2025-12-10 14:28:37.371990994 +0000 UTC m=+266.941208624" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.372284 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-ht7zw"] Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.372328 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.377113 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.378947 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.393051 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.393019786 podStartE2EDuration="23.393019786s" podCreationTimestamp="2025-12-10 14:28:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:28:37.386508753 +0000 UTC m=+266.955726383" watchObservedRunningTime="2025-12-10 14:28:37.393019786 +0000 UTC m=+266.962237456" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.417549 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.439371 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.452372 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.506058 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.803081 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.901068 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.940099 4847 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 14:28:37 crc kubenswrapper[4847]: I1210 14:28:37.961280 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.104637 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.122262 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.147087 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.172785 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.255319 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.392871 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.529995 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.564137 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.588241 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.661504 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.753291 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.765244 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" path="/var/lib/kubelet/pods/2e6718b4-c798-4e9a-9dff-fd3bcd9566db/volumes" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.769499 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.777226 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.807384 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.883570 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 14:28:38 crc kubenswrapper[4847]: I1210 14:28:38.986612 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.237221 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.256972 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.288394 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.491398 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.565086 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.567046 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.575789 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.605347 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.854581 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 14:28:39 crc kubenswrapper[4847]: I1210 14:28:39.984867 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.246535 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.318433 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.368551 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.410559 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.423566 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.495094 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.539325 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.590982 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.748691 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.954162 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 14:28:40 crc kubenswrapper[4847]: I1210 14:28:40.982853 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.019222 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.023752 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.143560 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.200822 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.377654 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.415570 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.504604 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.534364 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 14:28:41 crc kubenswrapper[4847]: I1210 14:28:41.537008 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.066843 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.079526 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.167442 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.207579 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.306164 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.490604 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.500842 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.544998 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-lkrpd"] Dec 10 14:28:42 crc kubenswrapper[4847]: E1210 14:28:42.545193 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" containerName="installer" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.545204 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" containerName="installer" Dec 10 14:28:42 crc kubenswrapper[4847]: E1210 14:28:42.545221 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" containerName="oauth-openshift" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.545227 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" containerName="oauth-openshift" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.545311 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a782016b-f55c-4dad-9617-19a0d4f0f42e" containerName="installer" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.545321 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6718b4-c798-4e9a-9dff-fd3bcd9566db" containerName="oauth-openshift" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.545697 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.548024 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.548509 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.548655 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.548911 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.549271 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.549893 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.550009 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.550022 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.549924 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.551748 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.551791 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.552001 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.556011 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.561550 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-lkrpd"] Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.565557 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.565910 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.677915 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678171 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678241 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678298 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678325 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678350 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678383 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hldc9\" (UniqueName: \"kubernetes.io/projected/8836770f-40d7-4af0-9e0d-cc12c7910c83-kube-api-access-hldc9\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678408 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678430 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678465 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-dir\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678500 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-policies\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678524 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678540 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.678566 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.712507 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.738320 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779756 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779791 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779817 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779848 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779864 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779882 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779911 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779928 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779950 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779974 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hldc9\" (UniqueName: \"kubernetes.io/projected/8836770f-40d7-4af0-9e0d-cc12c7910c83-kube-api-access-hldc9\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.779993 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780010 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780028 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-dir\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780042 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-policies\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780783 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-policies\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780833 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8836770f-40d7-4af0-9e0d-cc12c7910c83-audit-dir\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780841 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.780944 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.781173 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.785279 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.785611 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.785695 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.785742 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.786174 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.787080 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.792874 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.794023 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8836770f-40d7-4af0-9e0d-cc12c7910c83-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.794733 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.797703 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hldc9\" (UniqueName: \"kubernetes.io/projected/8836770f-40d7-4af0-9e0d-cc12c7910c83-kube-api-access-hldc9\") pod \"oauth-openshift-76766fc778-lkrpd\" (UID: \"8836770f-40d7-4af0-9e0d-cc12c7910c83\") " pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.861572 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:42 crc kubenswrapper[4847]: I1210 14:28:42.907444 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.040048 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.047520 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-lkrpd"] Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.151200 4847 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.155020 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.228028 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.241735 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" event={"ID":"8836770f-40d7-4af0-9e0d-cc12c7910c83","Type":"ContainerStarted","Data":"917c0c6c635dc8bbce1472840c37ebad26ad46c84459f3181ed37a4ca3e8804a"} Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.477373 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 14:28:43 crc kubenswrapper[4847]: I1210 14:28:43.625217 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.072252 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.249134 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" event={"ID":"8836770f-40d7-4af0-9e0d-cc12c7910c83","Type":"ContainerStarted","Data":"924aaac006d33ff21f25580118451ac1b00ec5519828f8ab4ca7c63521bee7a3"} Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.249913 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.253950 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.263815 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 14:28:44 crc kubenswrapper[4847]: I1210 14:28:44.267536 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-76766fc778-lkrpd" podStartSLOduration=55.267520017 podStartE2EDuration="55.267520017s" podCreationTimestamp="2025-12-10 14:27:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:28:44.265513197 +0000 UTC m=+273.834730837" watchObservedRunningTime="2025-12-10 14:28:44.267520017 +0000 UTC m=+273.836737647" Dec 10 14:28:46 crc kubenswrapper[4847]: I1210 14:28:46.269813 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 14:28:47 crc kubenswrapper[4847]: I1210 14:28:47.487579 4847 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 14:28:47 crc kubenswrapper[4847]: I1210 14:28:47.488227 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb" gracePeriod=5 Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.197127 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.197438 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.307118 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.307163 4847 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb" exitCode=137 Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.307204 4847 scope.go:117] "RemoveContainer" containerID="a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.307246 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311693 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311774 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311802 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311822 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311844 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311863 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311877 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.311902 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.312069 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.312197 4847 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.312215 4847 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.312257 4847 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.312281 4847 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.327455 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.329656 4847 scope.go:117] "RemoveContainer" containerID="a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb" Dec 10 14:28:53 crc kubenswrapper[4847]: E1210 14:28:53.330264 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb\": container with ID starting with a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb not found: ID does not exist" containerID="a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.330301 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb"} err="failed to get container status \"a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb\": rpc error: code = NotFound desc = could not find container \"a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb\": container with ID starting with a637e50e2b3da877ae738680e22bca00de64055201c61872e5d8e92e6c044ceb not found: ID does not exist" Dec 10 14:28:53 crc kubenswrapper[4847]: I1210 14:28:53.413639 4847 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.768454 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.769104 4847 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.781818 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.781857 4847 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="86f6e8a1-0258-494e-8dbd-06f9e8fdf887" Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.786281 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 14:28:54 crc kubenswrapper[4847]: I1210 14:28:54.786313 4847 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="86f6e8a1-0258-494e-8dbd-06f9e8fdf887" Dec 10 14:29:04 crc kubenswrapper[4847]: I1210 14:29:04.369655 4847 generic.go:334] "Generic (PLEG): container finished" podID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerID="ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa" exitCode=0 Dec 10 14:29:04 crc kubenswrapper[4847]: I1210 14:29:04.369778 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerDied","Data":"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa"} Dec 10 14:29:04 crc kubenswrapper[4847]: I1210 14:29:04.370636 4847 scope.go:117] "RemoveContainer" containerID="ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa" Dec 10 14:29:05 crc kubenswrapper[4847]: I1210 14:29:05.376957 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerStarted","Data":"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476"} Dec 10 14:29:05 crc kubenswrapper[4847]: I1210 14:29:05.377596 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:29:05 crc kubenswrapper[4847]: I1210 14:29:05.379860 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:29:10 crc kubenswrapper[4847]: I1210 14:29:10.622330 4847 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 10 14:29:25 crc kubenswrapper[4847]: I1210 14:29:25.721527 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:29:25 crc kubenswrapper[4847]: I1210 14:29:25.722370 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" podUID="743fcc8f-13eb-4ed8-a811-25ea265076bb" containerName="controller-manager" containerID="cri-o://8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e" gracePeriod=30 Dec 10 14:29:25 crc kubenswrapper[4847]: I1210 14:29:25.807449 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:29:25 crc kubenswrapper[4847]: I1210 14:29:25.807669 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" podUID="598e635e-4f60-4518-9021-456857be17e6" containerName="route-controller-manager" containerID="cri-o://8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963" gracePeriod=30 Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.141330 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.162937 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.251537 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert\") pod \"598e635e-4f60-4518-9021-456857be17e6\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.251688 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5jsc\" (UniqueName: \"kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc\") pod \"743fcc8f-13eb-4ed8-a811-25ea265076bb\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.252954 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles\") pod \"743fcc8f-13eb-4ed8-a811-25ea265076bb\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253054 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca\") pod \"598e635e-4f60-4518-9021-456857be17e6\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253098 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv5z5\" (UniqueName: \"kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5\") pod \"598e635e-4f60-4518-9021-456857be17e6\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253183 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config\") pod \"743fcc8f-13eb-4ed8-a811-25ea265076bb\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253274 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config\") pod \"598e635e-4f60-4518-9021-456857be17e6\" (UID: \"598e635e-4f60-4518-9021-456857be17e6\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253302 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert\") pod \"743fcc8f-13eb-4ed8-a811-25ea265076bb\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253328 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca\") pod \"743fcc8f-13eb-4ed8-a811-25ea265076bb\" (UID: \"743fcc8f-13eb-4ed8-a811-25ea265076bb\") " Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253572 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "743fcc8f-13eb-4ed8-a811-25ea265076bb" (UID: "743fcc8f-13eb-4ed8-a811-25ea265076bb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.253997 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.254033 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca" (OuterVolumeSpecName: "client-ca") pod "598e635e-4f60-4518-9021-456857be17e6" (UID: "598e635e-4f60-4518-9021-456857be17e6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.254068 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config" (OuterVolumeSpecName: "config") pod "598e635e-4f60-4518-9021-456857be17e6" (UID: "598e635e-4f60-4518-9021-456857be17e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.254548 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config" (OuterVolumeSpecName: "config") pod "743fcc8f-13eb-4ed8-a811-25ea265076bb" (UID: "743fcc8f-13eb-4ed8-a811-25ea265076bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.254591 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca" (OuterVolumeSpecName: "client-ca") pod "743fcc8f-13eb-4ed8-a811-25ea265076bb" (UID: "743fcc8f-13eb-4ed8-a811-25ea265076bb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.259627 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5" (OuterVolumeSpecName: "kube-api-access-bv5z5") pod "598e635e-4f60-4518-9021-456857be17e6" (UID: "598e635e-4f60-4518-9021-456857be17e6"). InnerVolumeSpecName "kube-api-access-bv5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.259650 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "743fcc8f-13eb-4ed8-a811-25ea265076bb" (UID: "743fcc8f-13eb-4ed8-a811-25ea265076bb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.259637 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "598e635e-4f60-4518-9021-456857be17e6" (UID: "598e635e-4f60-4518-9021-456857be17e6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.259747 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc" (OuterVolumeSpecName: "kube-api-access-p5jsc") pod "743fcc8f-13eb-4ed8-a811-25ea265076bb" (UID: "743fcc8f-13eb-4ed8-a811-25ea265076bb"). InnerVolumeSpecName "kube-api-access-p5jsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355304 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/598e635e-4f60-4518-9021-456857be17e6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355354 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5jsc\" (UniqueName: \"kubernetes.io/projected/743fcc8f-13eb-4ed8-a811-25ea265076bb-kube-api-access-p5jsc\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355364 4847 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355373 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv5z5\" (UniqueName: \"kubernetes.io/projected/598e635e-4f60-4518-9021-456857be17e6-kube-api-access-bv5z5\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355382 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355390 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598e635e-4f60-4518-9021-456857be17e6-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355398 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/743fcc8f-13eb-4ed8-a811-25ea265076bb-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.355405 4847 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/743fcc8f-13eb-4ed8-a811-25ea265076bb-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.496594 4847 generic.go:334] "Generic (PLEG): container finished" podID="743fcc8f-13eb-4ed8-a811-25ea265076bb" containerID="8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e" exitCode=0 Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.496666 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" event={"ID":"743fcc8f-13eb-4ed8-a811-25ea265076bb","Type":"ContainerDied","Data":"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e"} Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.496694 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" event={"ID":"743fcc8f-13eb-4ed8-a811-25ea265076bb","Type":"ContainerDied","Data":"af1ddd3dccbe89826ca21ca767cef10f16f4b331ed51ec66c8612a6d0feb1aa4"} Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.496728 4847 scope.go:117] "RemoveContainer" containerID="8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.496730 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5xqhd" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.500278 4847 generic.go:334] "Generic (PLEG): container finished" podID="598e635e-4f60-4518-9021-456857be17e6" containerID="8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963" exitCode=0 Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.500340 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.500669 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" event={"ID":"598e635e-4f60-4518-9021-456857be17e6","Type":"ContainerDied","Data":"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963"} Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.500748 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd" event={"ID":"598e635e-4f60-4518-9021-456857be17e6","Type":"ContainerDied","Data":"d164c32a972d75c113ad2f222624920c8d519e4f61a82d616a40f1c82450ef43"} Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.520988 4847 scope.go:117] "RemoveContainer" containerID="8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e" Dec 10 14:29:26 crc kubenswrapper[4847]: E1210 14:29:26.521800 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e\": container with ID starting with 8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e not found: ID does not exist" containerID="8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.521845 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e"} err="failed to get container status \"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e\": rpc error: code = NotFound desc = could not find container \"8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e\": container with ID starting with 8032abac7ed3118fc45d1742025f0bb6bfed5af6de87a91b4b447a686390ba9e not found: ID does not exist" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.521877 4847 scope.go:117] "RemoveContainer" containerID="8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.535948 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.542669 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-b7rjd"] Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.546587 4847 scope.go:117] "RemoveContainer" containerID="8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963" Dec 10 14:29:26 crc kubenswrapper[4847]: E1210 14:29:26.547121 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963\": container with ID starting with 8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963 not found: ID does not exist" containerID="8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.547159 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963"} err="failed to get container status \"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963\": rpc error: code = NotFound desc = could not find container \"8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963\": container with ID starting with 8175e7356470f0b0f71d96dd106a4a55a684836dc309520bf5482e963e18c963 not found: ID does not exist" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.547836 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.550940 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5xqhd"] Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.765622 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="598e635e-4f60-4518-9021-456857be17e6" path="/var/lib/kubelet/pods/598e635e-4f60-4518-9021-456857be17e6/volumes" Dec 10 14:29:26 crc kubenswrapper[4847]: I1210 14:29:26.766435 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743fcc8f-13eb-4ed8-a811-25ea265076bb" path="/var/lib/kubelet/pods/743fcc8f-13eb-4ed8-a811-25ea265076bb/volumes" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148247 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:29:27 crc kubenswrapper[4847]: E1210 14:29:27.148508 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743fcc8f-13eb-4ed8-a811-25ea265076bb" containerName="controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148562 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="743fcc8f-13eb-4ed8-a811-25ea265076bb" containerName="controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: E1210 14:29:27.148582 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598e635e-4f60-4518-9021-456857be17e6" containerName="route-controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148589 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="598e635e-4f60-4518-9021-456857be17e6" containerName="route-controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: E1210 14:29:27.148609 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148617 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148733 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="743fcc8f-13eb-4ed8-a811-25ea265076bb" containerName="controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148757 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="598e635e-4f60-4518-9021-456857be17e6" containerName="route-controller-manager" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.148772 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.149195 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.155454 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.155673 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.155466 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.156632 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.156743 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.156914 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.159604 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.160556 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b"] Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.161256 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.162673 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.164923 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b"] Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.165742 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.166004 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.166137 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.166254 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.166354 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.167801 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.268964 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-config\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269057 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269075 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd6c0295-551e-48ee-9cbb-501574520c0f-serving-cert\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269136 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-client-ca\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269228 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6c6k\" (UniqueName: \"kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269262 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269389 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269484 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.269554 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tvtd\" (UniqueName: \"kubernetes.io/projected/dd6c0295-551e-48ee-9cbb-501574520c0f-kube-api-access-7tvtd\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.371339 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.371894 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6c6k\" (UniqueName: \"kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.372250 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.372522 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.372678 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tvtd\" (UniqueName: \"kubernetes.io/projected/dd6c0295-551e-48ee-9cbb-501574520c0f-kube-api-access-7tvtd\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.372895 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-config\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.373066 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.373212 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd6c0295-551e-48ee-9cbb-501574520c0f-serving-cert\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.373436 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-client-ca\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.373520 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.374374 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.374424 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-client-ca\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.374608 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.374740 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd6c0295-551e-48ee-9cbb-501574520c0f-config\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.385035 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.385247 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd6c0295-551e-48ee-9cbb-501574520c0f-serving-cert\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.390175 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tvtd\" (UniqueName: \"kubernetes.io/projected/dd6c0295-551e-48ee-9cbb-501574520c0f-kube-api-access-7tvtd\") pod \"route-controller-manager-56874d89df-vp54b\" (UID: \"dd6c0295-551e-48ee-9cbb-501574520c0f\") " pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.392573 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6c6k\" (UniqueName: \"kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k\") pod \"controller-manager-856f74b64f-6d4cf\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.527177 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.549768 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.740966 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:29:27 crc kubenswrapper[4847]: W1210 14:29:27.752356 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fbdea7b_8301_4db5_bde7_ec4d7df47435.slice/crio-71244a80e018da09ad0451c10b740593bc610c882cf793205915db67c4629609 WatchSource:0}: Error finding container 71244a80e018da09ad0451c10b740593bc610c882cf793205915db67c4629609: Status 404 returned error can't find the container with id 71244a80e018da09ad0451c10b740593bc610c882cf793205915db67c4629609 Dec 10 14:29:27 crc kubenswrapper[4847]: I1210 14:29:27.788586 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b"] Dec 10 14:29:27 crc kubenswrapper[4847]: W1210 14:29:27.791960 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd6c0295_551e_48ee_9cbb_501574520c0f.slice/crio-5624a7e515b68c0ec6d4e1b524721da91c6e5bba47f125ef8a570823a4fc8a08 WatchSource:0}: Error finding container 5624a7e515b68c0ec6d4e1b524721da91c6e5bba47f125ef8a570823a4fc8a08: Status 404 returned error can't find the container with id 5624a7e515b68c0ec6d4e1b524721da91c6e5bba47f125ef8a570823a4fc8a08 Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.524440 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" event={"ID":"9fbdea7b-8301-4db5-bde7-ec4d7df47435","Type":"ContainerStarted","Data":"9a5e8277b5acc990800c769ba2735e6c31c5afe94e06d2098894ac805e152cf4"} Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.524786 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.524801 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" event={"ID":"9fbdea7b-8301-4db5-bde7-ec4d7df47435","Type":"ContainerStarted","Data":"71244a80e018da09ad0451c10b740593bc610c882cf793205915db67c4629609"} Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.526051 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" event={"ID":"dd6c0295-551e-48ee-9cbb-501574520c0f","Type":"ContainerStarted","Data":"57511e565fd7052f4f047b1dbdd5e36af1199311fce2622d9784486f78e11662"} Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.526080 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" event={"ID":"dd6c0295-551e-48ee-9cbb-501574520c0f","Type":"ContainerStarted","Data":"5624a7e515b68c0ec6d4e1b524721da91c6e5bba47f125ef8a570823a4fc8a08"} Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.526317 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.529122 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.530380 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.542550 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" podStartSLOduration=3.542531548 podStartE2EDuration="3.542531548s" podCreationTimestamp="2025-12-10 14:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:29:28.542319652 +0000 UTC m=+318.111537282" watchObservedRunningTime="2025-12-10 14:29:28.542531548 +0000 UTC m=+318.111749178" Dec 10 14:29:28 crc kubenswrapper[4847]: I1210 14:29:28.556534 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56874d89df-vp54b" podStartSLOduration=3.556514862 podStartE2EDuration="3.556514862s" podCreationTimestamp="2025-12-10 14:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:29:28.55575596 +0000 UTC m=+318.124973610" watchObservedRunningTime="2025-12-10 14:29:28.556514862 +0000 UTC m=+318.125732492" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.166544 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l"] Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.167922 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.170368 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.170883 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.179534 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l"] Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.205395 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.205471 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.205497 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dggwx\" (UniqueName: \"kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.307004 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.307090 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.307122 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dggwx\" (UniqueName: \"kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.308016 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.316408 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.323962 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dggwx\" (UniqueName: \"kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx\") pod \"collect-profiles-29422950-lmb8l\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.486013 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:00 crc kubenswrapper[4847]: I1210 14:30:00.893989 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l"] Dec 10 14:30:01 crc kubenswrapper[4847]: I1210 14:30:01.010671 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:30:01 crc kubenswrapper[4847]: I1210 14:30:01.010743 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:30:01 crc kubenswrapper[4847]: I1210 14:30:01.787882 4847 generic.go:334] "Generic (PLEG): container finished" podID="6c5545dc-5f70-4f7b-950d-077adbf83511" containerID="f5488e724c71f37aa03fa183d4c3a16aa08eedd692bbffe68f5d5c47606cd00d" exitCode=0 Dec 10 14:30:01 crc kubenswrapper[4847]: I1210 14:30:01.787929 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" event={"ID":"6c5545dc-5f70-4f7b-950d-077adbf83511","Type":"ContainerDied","Data":"f5488e724c71f37aa03fa183d4c3a16aa08eedd692bbffe68f5d5c47606cd00d"} Dec 10 14:30:01 crc kubenswrapper[4847]: I1210 14:30:01.788124 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" event={"ID":"6c5545dc-5f70-4f7b-950d-077adbf83511","Type":"ContainerStarted","Data":"aa2cb920bdaf6149ebb405d56e319016621e489439311546cb61c154756616fe"} Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.100209 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.144783 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume\") pod \"6c5545dc-5f70-4f7b-950d-077adbf83511\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.144915 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dggwx\" (UniqueName: \"kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx\") pod \"6c5545dc-5f70-4f7b-950d-077adbf83511\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.144955 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume\") pod \"6c5545dc-5f70-4f7b-950d-077adbf83511\" (UID: \"6c5545dc-5f70-4f7b-950d-077adbf83511\") " Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.148068 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume" (OuterVolumeSpecName: "config-volume") pod "6c5545dc-5f70-4f7b-950d-077adbf83511" (UID: "6c5545dc-5f70-4f7b-950d-077adbf83511"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.150686 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx" (OuterVolumeSpecName: "kube-api-access-dggwx") pod "6c5545dc-5f70-4f7b-950d-077adbf83511" (UID: "6c5545dc-5f70-4f7b-950d-077adbf83511"). InnerVolumeSpecName "kube-api-access-dggwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.150916 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6c5545dc-5f70-4f7b-950d-077adbf83511" (UID: "6c5545dc-5f70-4f7b-950d-077adbf83511"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.246760 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c5545dc-5f70-4f7b-950d-077adbf83511-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.246799 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dggwx\" (UniqueName: \"kubernetes.io/projected/6c5545dc-5f70-4f7b-950d-077adbf83511-kube-api-access-dggwx\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.246811 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c5545dc-5f70-4f7b-950d-077adbf83511-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.799522 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" event={"ID":"6c5545dc-5f70-4f7b-950d-077adbf83511","Type":"ContainerDied","Data":"aa2cb920bdaf6149ebb405d56e319016621e489439311546cb61c154756616fe"} Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.799565 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l" Dec 10 14:30:03 crc kubenswrapper[4847]: I1210 14:30:03.799568 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa2cb920bdaf6149ebb405d56e319016621e489439311546cb61c154756616fe" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.588810 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gpnxz"] Dec 10 14:30:09 crc kubenswrapper[4847]: E1210 14:30:09.589877 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5545dc-5f70-4f7b-950d-077adbf83511" containerName="collect-profiles" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.589890 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5545dc-5f70-4f7b-950d-077adbf83511" containerName="collect-profiles" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.589982 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c5545dc-5f70-4f7b-950d-077adbf83511" containerName="collect-profiles" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.590344 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.605457 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gpnxz"] Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625413 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-tls\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625487 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-certificates\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625521 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-bound-sa-token\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625548 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625578 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfxvs\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-kube-api-access-sfxvs\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625606 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625634 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-trusted-ca\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.625652 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.653397 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727328 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-certificates\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727401 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-bound-sa-token\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727439 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727483 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfxvs\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-kube-api-access-sfxvs\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727513 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-trusted-ca\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727536 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.727575 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-tls\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.728573 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.728645 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-certificates\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.728697 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-trusted-ca\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.734931 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.736203 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-registry-tls\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.745454 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfxvs\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-kube-api-access-sfxvs\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.745469 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9b5d8ba9-c7b5-4672-a956-78e68aa0d700-bound-sa-token\") pod \"image-registry-66df7c8f76-gpnxz\" (UID: \"9b5d8ba9-c7b5-4672-a956-78e68aa0d700\") " pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:09 crc kubenswrapper[4847]: I1210 14:30:09.941887 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:10 crc kubenswrapper[4847]: I1210 14:30:10.134541 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-gpnxz"] Dec 10 14:30:10 crc kubenswrapper[4847]: I1210 14:30:10.835847 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" event={"ID":"9b5d8ba9-c7b5-4672-a956-78e68aa0d700","Type":"ContainerStarted","Data":"c3a9cab3576351c34c4e74cce3e3bf72a9d5dd1398c10091dc848e5d5d4b5ded"} Dec 10 14:30:10 crc kubenswrapper[4847]: I1210 14:30:10.837429 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:10 crc kubenswrapper[4847]: I1210 14:30:10.837478 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" event={"ID":"9b5d8ba9-c7b5-4672-a956-78e68aa0d700","Type":"ContainerStarted","Data":"0ba2971674fa1d470230cb9fb1df37a637594a6650c26498e2f0119c30482538"} Dec 10 14:30:10 crc kubenswrapper[4847]: I1210 14:30:10.858967 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" podStartSLOduration=1.858945979 podStartE2EDuration="1.858945979s" podCreationTimestamp="2025-12-10 14:30:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:30:10.851680018 +0000 UTC m=+360.420897658" watchObservedRunningTime="2025-12-10 14:30:10.858945979 +0000 UTC m=+360.428163609" Dec 10 14:30:25 crc kubenswrapper[4847]: I1210 14:30:25.699789 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:30:25 crc kubenswrapper[4847]: I1210 14:30:25.700619 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" podUID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" containerName="controller-manager" containerID="cri-o://9a5e8277b5acc990800c769ba2735e6c31c5afe94e06d2098894ac805e152cf4" gracePeriod=30 Dec 10 14:30:25 crc kubenswrapper[4847]: I1210 14:30:25.921755 4847 generic.go:334] "Generic (PLEG): container finished" podID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" containerID="9a5e8277b5acc990800c769ba2735e6c31c5afe94e06d2098894ac805e152cf4" exitCode=0 Dec 10 14:30:25 crc kubenswrapper[4847]: I1210 14:30:25.921794 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" event={"ID":"9fbdea7b-8301-4db5-bde7-ec4d7df47435","Type":"ContainerDied","Data":"9a5e8277b5acc990800c769ba2735e6c31c5afe94e06d2098894ac805e152cf4"} Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.113828 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.149981 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca\") pod \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150031 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert\") pod \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150061 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles\") pod \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150089 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config\") pod \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150114 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6c6k\" (UniqueName: \"kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k\") pod \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\" (UID: \"9fbdea7b-8301-4db5-bde7-ec4d7df47435\") " Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150847 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9fbdea7b-8301-4db5-bde7-ec4d7df47435" (UID: "9fbdea7b-8301-4db5-bde7-ec4d7df47435"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150928 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca" (OuterVolumeSpecName: "client-ca") pod "9fbdea7b-8301-4db5-bde7-ec4d7df47435" (UID: "9fbdea7b-8301-4db5-bde7-ec4d7df47435"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.150934 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config" (OuterVolumeSpecName: "config") pod "9fbdea7b-8301-4db5-bde7-ec4d7df47435" (UID: "9fbdea7b-8301-4db5-bde7-ec4d7df47435"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.155257 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9fbdea7b-8301-4db5-bde7-ec4d7df47435" (UID: "9fbdea7b-8301-4db5-bde7-ec4d7df47435"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.155327 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k" (OuterVolumeSpecName: "kube-api-access-x6c6k") pod "9fbdea7b-8301-4db5-bde7-ec4d7df47435" (UID: "9fbdea7b-8301-4db5-bde7-ec4d7df47435"). InnerVolumeSpecName "kube-api-access-x6c6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.251884 4847 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.251922 4847 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fbdea7b-8301-4db5-bde7-ec4d7df47435-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.251933 4847 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.251949 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fbdea7b-8301-4db5-bde7-ec4d7df47435-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.251960 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6c6k\" (UniqueName: \"kubernetes.io/projected/9fbdea7b-8301-4db5-bde7-ec4d7df47435-kube-api-access-x6c6k\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.928650 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" event={"ID":"9fbdea7b-8301-4db5-bde7-ec4d7df47435","Type":"ContainerDied","Data":"71244a80e018da09ad0451c10b740593bc610c882cf793205915db67c4629609"} Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.928738 4847 scope.go:117] "RemoveContainer" containerID="9a5e8277b5acc990800c769ba2735e6c31c5afe94e06d2098894ac805e152cf4" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.928846 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-856f74b64f-6d4cf" Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.946329 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:30:26 crc kubenswrapper[4847]: I1210 14:30:26.949324 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-856f74b64f-6d4cf"] Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.183002 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-d75cffd9d-xvgwt"] Dec 10 14:30:27 crc kubenswrapper[4847]: E1210 14:30:27.183212 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" containerName="controller-manager" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.183224 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" containerName="controller-manager" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.183345 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" containerName="controller-manager" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.183781 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.185099 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.185305 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.186378 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.186640 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.186780 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.187423 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.192027 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.194742 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d75cffd9d-xvgwt"] Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.264247 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z68kv\" (UniqueName: \"kubernetes.io/projected/6a4ac791-24b4-4667-91a4-dfe96efd402d-kube-api-access-z68kv\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.264342 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-client-ca\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.264381 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-config\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.264424 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-proxy-ca-bundles\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.264452 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a4ac791-24b4-4667-91a4-dfe96efd402d-serving-cert\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.365245 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z68kv\" (UniqueName: \"kubernetes.io/projected/6a4ac791-24b4-4667-91a4-dfe96efd402d-kube-api-access-z68kv\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.365324 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-client-ca\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.365369 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-config\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.365412 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-proxy-ca-bundles\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.365441 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a4ac791-24b4-4667-91a4-dfe96efd402d-serving-cert\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.367303 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-client-ca\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.367737 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-config\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.367963 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a4ac791-24b4-4667-91a4-dfe96efd402d-proxy-ca-bundles\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.371263 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a4ac791-24b4-4667-91a4-dfe96efd402d-serving-cert\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.383589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z68kv\" (UniqueName: \"kubernetes.io/projected/6a4ac791-24b4-4667-91a4-dfe96efd402d-kube-api-access-z68kv\") pod \"controller-manager-d75cffd9d-xvgwt\" (UID: \"6a4ac791-24b4-4667-91a4-dfe96efd402d\") " pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.511841 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.701499 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-d75cffd9d-xvgwt"] Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.934559 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" event={"ID":"6a4ac791-24b4-4667-91a4-dfe96efd402d","Type":"ContainerStarted","Data":"9b4075652c3539888ab7768fb11b3efcdd83c99f3ee4995b2ea44a47285ffb3a"} Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.934604 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" event={"ID":"6a4ac791-24b4-4667-91a4-dfe96efd402d","Type":"ContainerStarted","Data":"e06f329eda3ca91141f9c0828e44dffdb38aca9e2aeee0bb9a8ca4285c4357cb"} Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.934927 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.939577 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" Dec 10 14:30:27 crc kubenswrapper[4847]: I1210 14:30:27.980463 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-d75cffd9d-xvgwt" podStartSLOduration=2.98044577 podStartE2EDuration="2.98044577s" podCreationTimestamp="2025-12-10 14:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:30:27.951404405 +0000 UTC m=+377.520622045" watchObservedRunningTime="2025-12-10 14:30:27.98044577 +0000 UTC m=+377.549663400" Dec 10 14:30:28 crc kubenswrapper[4847]: I1210 14:30:28.769082 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fbdea7b-8301-4db5-bde7-ec4d7df47435" path="/var/lib/kubelet/pods/9fbdea7b-8301-4db5-bde7-ec4d7df47435/volumes" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.001367 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.001749 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kvfs8" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="registry-server" containerID="cri-o://79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29" gracePeriod=30 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.013519 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.013854 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sm78p" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="registry-server" containerID="cri-o://6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1" gracePeriod=30 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.021226 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.021461 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" containerID="cri-o://7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476" gracePeriod=30 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.036472 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.037063 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4tsg4" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="registry-server" containerID="cri-o://c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906" gracePeriod=30 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.055064 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dz9r4"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.056911 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.070353 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.070621 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8gv4z" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="registry-server" containerID="cri-o://047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2" gracePeriod=30 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.076107 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dz9r4"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.086518 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.086590 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.086630 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwfjd\" (UniqueName: \"kubernetes.io/projected/f55386e3-acde-49bd-941e-a92591ec95de-kube-api-access-fwfjd\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.188500 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.188580 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.188624 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwfjd\" (UniqueName: \"kubernetes.io/projected/f55386e3-acde-49bd-941e-a92591ec95de-kube-api-access-fwfjd\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.190041 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.200168 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f55386e3-acde-49bd-941e-a92591ec95de-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.207532 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwfjd\" (UniqueName: \"kubernetes.io/projected/f55386e3-acde-49bd-941e-a92591ec95de-kube-api-access-fwfjd\") pod \"marketplace-operator-79b997595-dz9r4\" (UID: \"f55386e3-acde-49bd-941e-a92591ec95de\") " pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.377203 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.605310 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.612121 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.640577 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.643558 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.664064 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dz9r4"] Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696220 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities\") pod \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696291 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities\") pod \"d73279bb-7be5-4397-9aee-ac26a8861a58\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696319 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities\") pod \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696382 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content\") pod \"d73279bb-7be5-4397-9aee-ac26a8861a58\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696448 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9j64\" (UniqueName: \"kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64\") pod \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696482 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content\") pod \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\" (UID: \"ddfe3352-42db-46fd-9e40-b4f4bf5f3292\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696536 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klx7g\" (UniqueName: \"kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g\") pod \"a53cedc2-877b-4e7f-9440-710de3a2faa7\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696594 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content\") pod \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696668 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics\") pod \"a53cedc2-877b-4e7f-9440-710de3a2faa7\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696691 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksdhn\" (UniqueName: \"kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn\") pod \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\" (UID: \"3987f250-b286-4eed-86e0-2c9e6abd9e3f\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696846 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sq9b\" (UniqueName: \"kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b\") pod \"d73279bb-7be5-4397-9aee-ac26a8861a58\" (UID: \"d73279bb-7be5-4397-9aee-ac26a8861a58\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.696874 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca\") pod \"a53cedc2-877b-4e7f-9440-710de3a2faa7\" (UID: \"a53cedc2-877b-4e7f-9440-710de3a2faa7\") " Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.698377 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a53cedc2-877b-4e7f-9440-710de3a2faa7" (UID: "a53cedc2-877b-4e7f-9440-710de3a2faa7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.700475 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities" (OuterVolumeSpecName: "utilities") pod "ddfe3352-42db-46fd-9e40-b4f4bf5f3292" (UID: "ddfe3352-42db-46fd-9e40-b4f4bf5f3292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.701258 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities" (OuterVolumeSpecName: "utilities") pod "3987f250-b286-4eed-86e0-2c9e6abd9e3f" (UID: "3987f250-b286-4eed-86e0-2c9e6abd9e3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.702095 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities" (OuterVolumeSpecName: "utilities") pod "d73279bb-7be5-4397-9aee-ac26a8861a58" (UID: "d73279bb-7be5-4397-9aee-ac26a8861a58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.707817 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b" (OuterVolumeSpecName: "kube-api-access-6sq9b") pod "d73279bb-7be5-4397-9aee-ac26a8861a58" (UID: "d73279bb-7be5-4397-9aee-ac26a8861a58"). InnerVolumeSpecName "kube-api-access-6sq9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.709670 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g" (OuterVolumeSpecName: "kube-api-access-klx7g") pod "a53cedc2-877b-4e7f-9440-710de3a2faa7" (UID: "a53cedc2-877b-4e7f-9440-710de3a2faa7"). InnerVolumeSpecName "kube-api-access-klx7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.718445 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a53cedc2-877b-4e7f-9440-710de3a2faa7" (UID: "a53cedc2-877b-4e7f-9440-710de3a2faa7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.718840 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64" (OuterVolumeSpecName: "kube-api-access-f9j64") pod "ddfe3352-42db-46fd-9e40-b4f4bf5f3292" (UID: "ddfe3352-42db-46fd-9e40-b4f4bf5f3292"). InnerVolumeSpecName "kube-api-access-f9j64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.726565 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn" (OuterVolumeSpecName: "kube-api-access-ksdhn") pod "3987f250-b286-4eed-86e0-2c9e6abd9e3f" (UID: "3987f250-b286-4eed-86e0-2c9e6abd9e3f"). InnerVolumeSpecName "kube-api-access-ksdhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.747896 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ddfe3352-42db-46fd-9e40-b4f4bf5f3292" (UID: "ddfe3352-42db-46fd-9e40-b4f4bf5f3292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.771503 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3987f250-b286-4eed-86e0-2c9e6abd9e3f" (UID: "3987f250-b286-4eed-86e0-2c9e6abd9e3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798280 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798671 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798680 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798692 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9j64\" (UniqueName: \"kubernetes.io/projected/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-kube-api-access-f9j64\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798703 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ddfe3352-42db-46fd-9e40-b4f4bf5f3292-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798742 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klx7g\" (UniqueName: \"kubernetes.io/projected/a53cedc2-877b-4e7f-9440-710de3a2faa7-kube-api-access-klx7g\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798755 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3987f250-b286-4eed-86e0-2c9e6abd9e3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798768 4847 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798820 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksdhn\" (UniqueName: \"kubernetes.io/projected/3987f250-b286-4eed-86e0-2c9e6abd9e3f-kube-api-access-ksdhn\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798831 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sq9b\" (UniqueName: \"kubernetes.io/projected/d73279bb-7be5-4397-9aee-ac26a8861a58-kube-api-access-6sq9b\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.798840 4847 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a53cedc2-877b-4e7f-9440-710de3a2faa7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.847082 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d73279bb-7be5-4397-9aee-ac26a8861a58" (UID: "d73279bb-7be5-4397-9aee-ac26a8861a58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.899857 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d73279bb-7be5-4397-9aee-ac26a8861a58-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.908208 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.948642 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-gpnxz" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.961259 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" event={"ID":"f55386e3-acde-49bd-941e-a92591ec95de","Type":"ContainerStarted","Data":"072ee3f514dadc7e8c06980b7e22f3ef239fdf2b648e7f60735c80a3dd8891a7"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.961345 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" event={"ID":"f55386e3-acde-49bd-941e-a92591ec95de","Type":"ContainerStarted","Data":"b2801c9a4af65cbfbc32e1f3fa579e5b6ed8fba14d03a4586e3acf3e02caa6e0"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.961544 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.963475 4847 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dz9r4 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.963516 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" podUID="f55386e3-acde-49bd-941e-a92591ec95de" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.964828 4847 generic.go:334] "Generic (PLEG): container finished" podID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerID="6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1" exitCode=0 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.964899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerDied","Data":"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.964926 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sm78p" event={"ID":"3987f250-b286-4eed-86e0-2c9e6abd9e3f","Type":"ContainerDied","Data":"bbf70442d4795435f6ff4b4a5580a1b71eb5d1d941077a90b96e75f703c91624"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.964950 4847 scope.go:117] "RemoveContainer" containerID="6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.965059 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sm78p" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.974268 4847 generic.go:334] "Generic (PLEG): container finished" podID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerID="7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476" exitCode=0 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.974359 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerDied","Data":"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.974388 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" event={"ID":"a53cedc2-877b-4e7f-9440-710de3a2faa7","Type":"ContainerDied","Data":"b97d0fb9b31de057f4b3cd61eb4c5909dab0ca9060c308948a2df7bd4828659e"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.974444 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-98dk2" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.984497 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6f20b79-6bb4-4832-a607-88805371713c" containerID="79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29" exitCode=0 Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.984602 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerDied","Data":"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.984638 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvfs8" event={"ID":"a6f20b79-6bb4-4832-a607-88805371713c","Type":"ContainerDied","Data":"5594a43853cffe3910c58c15db46c0c81f55e34598d150080cd2c0a429a2dda8"} Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.984745 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvfs8" Dec 10 14:30:29 crc kubenswrapper[4847]: I1210 14:30:29.987936 4847 scope.go:117] "RemoveContainer" containerID="f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.005702 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" podStartSLOduration=1.005676595 podStartE2EDuration="1.005676595s" podCreationTimestamp="2025-12-10 14:30:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:30:30.000381328 +0000 UTC m=+379.569598968" watchObservedRunningTime="2025-12-10 14:30:30.005676595 +0000 UTC m=+379.574894225" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.010402 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.015395 4847 generic.go:334] "Generic (PLEG): container finished" podID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerID="c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906" exitCode=0 Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.015501 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerDied","Data":"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906"} Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.015536 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tsg4" event={"ID":"ddfe3352-42db-46fd-9e40-b4f4bf5f3292","Type":"ContainerDied","Data":"943de1cdf53a1370340d632953bae859d9f2ffe53e2b3c3ec4a1a0055fdc7b83"} Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.015610 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tsg4" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.021346 4847 generic.go:334] "Generic (PLEG): container finished" podID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerID="047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2" exitCode=0 Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.021443 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8gv4z" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.021491 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerDied","Data":"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2"} Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.021521 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8gv4z" event={"ID":"d73279bb-7be5-4397-9aee-ac26a8861a58","Type":"ContainerDied","Data":"4533312a682cfceb88fbfd162744b930d56ba1245e8a107912d91c3c1ee638bc"} Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.044800 4847 scope.go:117] "RemoveContainer" containerID="583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.075941 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.082169 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sm78p"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.090692 4847 scope.go:117] "RemoveContainer" containerID="6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.092420 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1\": container with ID starting with 6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1 not found: ID does not exist" containerID="6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.092485 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1"} err="failed to get container status \"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1\": rpc error: code = NotFound desc = could not find container \"6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1\": container with ID starting with 6692c2b16993b79527413fa69d04d97c9d109265e011167b9bb781e4405fbfc1 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.092531 4847 scope.go:117] "RemoveContainer" containerID="f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.092984 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1\": container with ID starting with f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1 not found: ID does not exist" containerID="f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.093028 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1"} err="failed to get container status \"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1\": rpc error: code = NotFound desc = could not find container \"f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1\": container with ID starting with f61f83df632d22ded647712acbe3d143e84d858d59b3dae9942cf55a2998bab1 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.093063 4847 scope.go:117] "RemoveContainer" containerID="583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.093586 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952\": container with ID starting with 583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952 not found: ID does not exist" containerID="583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.093657 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952"} err="failed to get container status \"583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952\": rpc error: code = NotFound desc = could not find container \"583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952\": container with ID starting with 583b55fb190f1fb9f84d4a39a5944873dfd3d4a9ea230b69daa1996119a3f952 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.093688 4847 scope.go:117] "RemoveContainer" containerID="7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.095981 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.101678 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj6vw\" (UniqueName: \"kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw\") pod \"a6f20b79-6bb4-4832-a607-88805371713c\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.101763 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities\") pod \"a6f20b79-6bb4-4832-a607-88805371713c\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.101856 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content\") pod \"a6f20b79-6bb4-4832-a607-88805371713c\" (UID: \"a6f20b79-6bb4-4832-a607-88805371713c\") " Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.105558 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities" (OuterVolumeSpecName: "utilities") pod "a6f20b79-6bb4-4832-a607-88805371713c" (UID: "a6f20b79-6bb4-4832-a607-88805371713c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.108698 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-98dk2"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.111836 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw" (OuterVolumeSpecName: "kube-api-access-wj6vw") pod "a6f20b79-6bb4-4832-a607-88805371713c" (UID: "a6f20b79-6bb4-4832-a607-88805371713c"). InnerVolumeSpecName "kube-api-access-wj6vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.113480 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.117332 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tsg4"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.120301 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.124372 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8gv4z"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.136147 4847 scope.go:117] "RemoveContainer" containerID="ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.149488 4847 scope.go:117] "RemoveContainer" containerID="7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.150808 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476\": container with ID starting with 7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476 not found: ID does not exist" containerID="7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.150843 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476"} err="failed to get container status \"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476\": rpc error: code = NotFound desc = could not find container \"7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476\": container with ID starting with 7647543a7af07ff6b504be7d264f9a5a88d23528c7a70f416cc12c1a3848e476 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.150879 4847 scope.go:117] "RemoveContainer" containerID="ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.151161 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa\": container with ID starting with ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa not found: ID does not exist" containerID="ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.151202 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa"} err="failed to get container status \"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa\": rpc error: code = NotFound desc = could not find container \"ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa\": container with ID starting with ae0b39eaa7ac49ee7128a8518bf6b7b8cd54ec1b44c969d455322fdb1c0c65aa not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.151228 4847 scope.go:117] "RemoveContainer" containerID="79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.162982 4847 scope.go:117] "RemoveContainer" containerID="b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.178012 4847 scope.go:117] "RemoveContainer" containerID="91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.199703 4847 scope.go:117] "RemoveContainer" containerID="79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.200447 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29\": container with ID starting with 79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29 not found: ID does not exist" containerID="79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.200498 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29"} err="failed to get container status \"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29\": rpc error: code = NotFound desc = could not find container \"79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29\": container with ID starting with 79ed878b82b115ffb99d6368eaed3844102b83cf5a10664a80c3a47a39487a29 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.200529 4847 scope.go:117] "RemoveContainer" containerID="b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.201021 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861\": container with ID starting with b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861 not found: ID does not exist" containerID="b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.201043 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861"} err="failed to get container status \"b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861\": rpc error: code = NotFound desc = could not find container \"b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861\": container with ID starting with b6e9ea91655a6bcf8be036f90745bfdad412da8200e3eca19fd7b5388d9c0861 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.201059 4847 scope.go:117] "RemoveContainer" containerID="91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.201290 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48\": container with ID starting with 91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48 not found: ID does not exist" containerID="91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.201311 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48"} err="failed to get container status \"91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48\": rpc error: code = NotFound desc = could not find container \"91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48\": container with ID starting with 91f87407199c3635a6d9f45e5c1c180ab951efe97db36497f3650486241eff48 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.201323 4847 scope.go:117] "RemoveContainer" containerID="c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.203465 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj6vw\" (UniqueName: \"kubernetes.io/projected/a6f20b79-6bb4-4832-a607-88805371713c-kube-api-access-wj6vw\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.203492 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.204934 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6f20b79-6bb4-4832-a607-88805371713c" (UID: "a6f20b79-6bb4-4832-a607-88805371713c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.223477 4847 scope.go:117] "RemoveContainer" containerID="11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.247352 4847 scope.go:117] "RemoveContainer" containerID="48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.264287 4847 scope.go:117] "RemoveContainer" containerID="c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.264968 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906\": container with ID starting with c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906 not found: ID does not exist" containerID="c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.265020 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906"} err="failed to get container status \"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906\": rpc error: code = NotFound desc = could not find container \"c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906\": container with ID starting with c2fcd6575381ba888657f497e514be3b2409e4dfd89929bb1b5718e64642e906 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.265048 4847 scope.go:117] "RemoveContainer" containerID="11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.265382 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019\": container with ID starting with 11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019 not found: ID does not exist" containerID="11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.265417 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019"} err="failed to get container status \"11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019\": rpc error: code = NotFound desc = could not find container \"11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019\": container with ID starting with 11cba1b388c6914d80bbfcdd7bc430771c700fb2b20ee77e1db258ff249eb019 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.265440 4847 scope.go:117] "RemoveContainer" containerID="48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.266068 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63\": container with ID starting with 48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63 not found: ID does not exist" containerID="48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.266092 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63"} err="failed to get container status \"48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63\": rpc error: code = NotFound desc = could not find container \"48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63\": container with ID starting with 48bfff72d6b5a8bdea0ef8d52a4d1faf7d30478afae931665b38082785bf1b63 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.266106 4847 scope.go:117] "RemoveContainer" containerID="047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.282499 4847 scope.go:117] "RemoveContainer" containerID="3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.297285 4847 scope.go:117] "RemoveContainer" containerID="4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.304821 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6f20b79-6bb4-4832-a607-88805371713c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.318662 4847 scope.go:117] "RemoveContainer" containerID="047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.320462 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2\": container with ID starting with 047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2 not found: ID does not exist" containerID="047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.320505 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2"} err="failed to get container status \"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2\": rpc error: code = NotFound desc = could not find container \"047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2\": container with ID starting with 047784d052fe28e5a5e45ec6c515c27b8609543595472169711c0b93e84d93e2 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.320539 4847 scope.go:117] "RemoveContainer" containerID="3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.320578 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.321463 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587\": container with ID starting with 3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587 not found: ID does not exist" containerID="3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.321492 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587"} err="failed to get container status \"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587\": rpc error: code = NotFound desc = could not find container \"3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587\": container with ID starting with 3e490e12a1a90619284a12d5c4aa02f69510907818809395c3f7abb78acd1587 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.321508 4847 scope.go:117] "RemoveContainer" containerID="4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378" Dec 10 14:30:30 crc kubenswrapper[4847]: E1210 14:30:30.321920 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378\": container with ID starting with 4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378 not found: ID does not exist" containerID="4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.321958 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378"} err="failed to get container status \"4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378\": rpc error: code = NotFound desc = could not find container \"4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378\": container with ID starting with 4f6ecf5a5c90f9a5b92fa53fcdba84aea9029fcb3d20dcea9a4fb835252df378 not found: ID does not exist" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.328424 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kvfs8"] Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.766434 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" path="/var/lib/kubelet/pods/3987f250-b286-4eed-86e0-2c9e6abd9e3f/volumes" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.767085 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" path="/var/lib/kubelet/pods/a53cedc2-877b-4e7f-9440-710de3a2faa7/volumes" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.767520 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6f20b79-6bb4-4832-a607-88805371713c" path="/var/lib/kubelet/pods/a6f20b79-6bb4-4832-a607-88805371713c/volumes" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.768088 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" path="/var/lib/kubelet/pods/d73279bb-7be5-4397-9aee-ac26a8861a58/volumes" Dec 10 14:30:30 crc kubenswrapper[4847]: I1210 14:30:30.768734 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" path="/var/lib/kubelet/pods/ddfe3352-42db-46fd-9e40-b4f4bf5f3292/volumes" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.010738 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.010791 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.038752 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dz9r4" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.199813 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4frbk"] Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200063 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200079 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200094 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200104 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200113 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200122 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200132 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200139 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200147 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200155 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200167 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200174 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200201 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200209 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200222 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200230 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200240 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200248 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200259 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200266 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200278 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200287 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200297 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200305 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="extract-utilities" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200313 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200320 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="extract-content" Dec 10 14:30:31 crc kubenswrapper[4847]: E1210 14:30:31.200328 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200335 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200445 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200460 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="3987f250-b286-4eed-86e0-2c9e6abd9e3f" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200475 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73279bb-7be5-4397-9aee-ac26a8861a58" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200487 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f20b79-6bb4-4832-a607-88805371713c" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200498 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddfe3352-42db-46fd-9e40-b4f4bf5f3292" containerName="registry-server" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.200686 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53cedc2-877b-4e7f-9440-710de3a2faa7" containerName="marketplace-operator" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.201298 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.206376 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.211747 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4frbk"] Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.319011 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-catalog-content\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.319112 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-utilities\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.319180 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdnx7\" (UniqueName: \"kubernetes.io/projected/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-kube-api-access-vdnx7\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.404830 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.406567 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.411058 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.412698 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.419799 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-utilities\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.419883 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdnx7\" (UniqueName: \"kubernetes.io/projected/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-kube-api-access-vdnx7\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.419916 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-catalog-content\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.420306 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-catalog-content\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.420513 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-utilities\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.438530 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdnx7\" (UniqueName: \"kubernetes.io/projected/ea23c9a6-79c4-48c1-bb39-2c75dc805ea8-kube-api-access-vdnx7\") pod \"redhat-marketplace-4frbk\" (UID: \"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8\") " pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.521472 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6927\" (UniqueName: \"kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.521957 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.522102 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.527218 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.624806 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.624912 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6927\" (UniqueName: \"kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.624953 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.625411 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.625531 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.646935 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6927\" (UniqueName: \"kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927\") pod \"redhat-operators-tc2pm\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.723326 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:31 crc kubenswrapper[4847]: I1210 14:30:31.940748 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4frbk"] Dec 10 14:30:31 crc kubenswrapper[4847]: W1210 14:30:31.948540 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea23c9a6_79c4_48c1_bb39_2c75dc805ea8.slice/crio-7ec5e78febea220911085005371edf4e08922341dc47c4ba437ba44b6811f76e WatchSource:0}: Error finding container 7ec5e78febea220911085005371edf4e08922341dc47c4ba437ba44b6811f76e: Status 404 returned error can't find the container with id 7ec5e78febea220911085005371edf4e08922341dc47c4ba437ba44b6811f76e Dec 10 14:30:32 crc kubenswrapper[4847]: I1210 14:30:32.040247 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4frbk" event={"ID":"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8","Type":"ContainerStarted","Data":"7ec5e78febea220911085005371edf4e08922341dc47c4ba437ba44b6811f76e"} Dec 10 14:30:32 crc kubenswrapper[4847]: I1210 14:30:32.136839 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:30:32 crc kubenswrapper[4847]: W1210 14:30:32.163491 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc69be8a_98d2_4ebe_a1b7_a2df888c7bbe.slice/crio-b9fa17489a2d2fdbbb112a6c7d8550bc1faf33b3bf6afe175656cd312db7c5a0 WatchSource:0}: Error finding container b9fa17489a2d2fdbbb112a6c7d8550bc1faf33b3bf6afe175656cd312db7c5a0: Status 404 returned error can't find the container with id b9fa17489a2d2fdbbb112a6c7d8550bc1faf33b3bf6afe175656cd312db7c5a0 Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.051224 4847 generic.go:334] "Generic (PLEG): container finished" podID="ea23c9a6-79c4-48c1-bb39-2c75dc805ea8" containerID="9df7ac47d25302a2e2c3ae7e0cb40244f663bf070851e8e85c9c20d38ae1644d" exitCode=0 Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.052828 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4frbk" event={"ID":"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8","Type":"ContainerDied","Data":"9df7ac47d25302a2e2c3ae7e0cb40244f663bf070851e8e85c9c20d38ae1644d"} Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.061034 4847 generic.go:334] "Generic (PLEG): container finished" podID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerID="2dad778a5c82fabc6cd9c7e488a77aaca4374e37e7f60880ba560c5481da0b4d" exitCode=0 Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.061078 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerDied","Data":"2dad778a5c82fabc6cd9c7e488a77aaca4374e37e7f60880ba560c5481da0b4d"} Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.061103 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerStarted","Data":"b9fa17489a2d2fdbbb112a6c7d8550bc1faf33b3bf6afe175656cd312db7c5a0"} Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.605172 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.606120 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.609566 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.619742 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.757779 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.757827 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.757932 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqfch\" (UniqueName: \"kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.802543 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.804954 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.808679 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.820949 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.860866 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqfch\" (UniqueName: \"kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.861087 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.861125 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.862510 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.863767 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.886305 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqfch\" (UniqueName: \"kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch\") pod \"certified-operators-s6wpx\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.930182 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.963040 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.963087 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npgts\" (UniqueName: \"kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:33 crc kubenswrapper[4847]: I1210 14:30:33.963140 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.064596 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.065251 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npgts\" (UniqueName: \"kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.065298 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.065307 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.065778 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.072335 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerStarted","Data":"40fe880cefdae4b028349b8568f3cbc966ff7af1c9aafe1b94f84db3994bb30f"} Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.078133 4847 generic.go:334] "Generic (PLEG): container finished" podID="ea23c9a6-79c4-48c1-bb39-2c75dc805ea8" containerID="9ee36e1cc9fa5e36a59eb6216ce9d4722c78b959fdbeecc0f29d5622cf5cceba" exitCode=0 Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.078226 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4frbk" event={"ID":"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8","Type":"ContainerDied","Data":"9ee36e1cc9fa5e36a59eb6216ce9d4722c78b959fdbeecc0f29d5622cf5cceba"} Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.081916 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npgts\" (UniqueName: \"kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts\") pod \"community-operators-r78kt\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.129305 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.340856 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 14:30:34 crc kubenswrapper[4847]: W1210 14:30:34.345665 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d91abd9_2ff0_4696_8942_32149e698707.slice/crio-329a5253d6d2842d980cf9f492fb5830fcb78425d9f9e71bfa6b9f0801ec2dab WatchSource:0}: Error finding container 329a5253d6d2842d980cf9f492fb5830fcb78425d9f9e71bfa6b9f0801ec2dab: Status 404 returned error can't find the container with id 329a5253d6d2842d980cf9f492fb5830fcb78425d9f9e71bfa6b9f0801ec2dab Dec 10 14:30:34 crc kubenswrapper[4847]: I1210 14:30:34.516135 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:30:34 crc kubenswrapper[4847]: W1210 14:30:34.617151 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68557631_c526_4bcf_8f5a_0f5b9e89a3f9.slice/crio-4491b236d50cf26108cea387cd75904aa9ddd2b60afe590925c6e245521f35f8 WatchSource:0}: Error finding container 4491b236d50cf26108cea387cd75904aa9ddd2b60afe590925c6e245521f35f8: Status 404 returned error can't find the container with id 4491b236d50cf26108cea387cd75904aa9ddd2b60afe590925c6e245521f35f8 Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.084081 4847 generic.go:334] "Generic (PLEG): container finished" podID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerID="40fe880cefdae4b028349b8568f3cbc966ff7af1c9aafe1b94f84db3994bb30f" exitCode=0 Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.084176 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerDied","Data":"40fe880cefdae4b028349b8568f3cbc966ff7af1c9aafe1b94f84db3994bb30f"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.086015 4847 generic.go:334] "Generic (PLEG): container finished" podID="0d91abd9-2ff0-4696-8942-32149e698707" containerID="366c087f642193fedbe6897b048b89acb3983aec4bccab1217a9a7156dde0718" exitCode=0 Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.086064 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerDied","Data":"366c087f642193fedbe6897b048b89acb3983aec4bccab1217a9a7156dde0718"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.086351 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerStarted","Data":"329a5253d6d2842d980cf9f492fb5830fcb78425d9f9e71bfa6b9f0801ec2dab"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.091435 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4frbk" event={"ID":"ea23c9a6-79c4-48c1-bb39-2c75dc805ea8","Type":"ContainerStarted","Data":"cb9caf8966624a50700e1ead070e92d21bd193623fc3c5986d86950f5cc0a5d9"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.093202 4847 generic.go:334] "Generic (PLEG): container finished" podID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerID="9ad7113fb685bd443625d6a3a2705366f5539129a48900acd4eb6189731f0fef" exitCode=0 Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.093251 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerDied","Data":"9ad7113fb685bd443625d6a3a2705366f5539129a48900acd4eb6189731f0fef"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.093274 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerStarted","Data":"4491b236d50cf26108cea387cd75904aa9ddd2b60afe590925c6e245521f35f8"} Dec 10 14:30:35 crc kubenswrapper[4847]: I1210 14:30:35.150743 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4frbk" podStartSLOduration=2.561876979 podStartE2EDuration="4.150724223s" podCreationTimestamp="2025-12-10 14:30:31 +0000 UTC" firstStartedPulling="2025-12-10 14:30:33.056063774 +0000 UTC m=+382.625281404" lastFinishedPulling="2025-12-10 14:30:34.644911018 +0000 UTC m=+384.214128648" observedRunningTime="2025-12-10 14:30:35.148268894 +0000 UTC m=+384.717486544" watchObservedRunningTime="2025-12-10 14:30:35.150724223 +0000 UTC m=+384.719941853" Dec 10 14:30:36 crc kubenswrapper[4847]: I1210 14:30:36.101011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerStarted","Data":"87f1d6aea560eba1ef453be09d695c66b811d04d13f70693cfa32579bf0926cf"} Dec 10 14:30:36 crc kubenswrapper[4847]: I1210 14:30:36.103086 4847 generic.go:334] "Generic (PLEG): container finished" podID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerID="7cf7549138c29660d9cf9f62372db2f96746ff66efa71003242f0f6559f4e2b8" exitCode=0 Dec 10 14:30:36 crc kubenswrapper[4847]: I1210 14:30:36.103593 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerDied","Data":"7cf7549138c29660d9cf9f62372db2f96746ff66efa71003242f0f6559f4e2b8"} Dec 10 14:30:37 crc kubenswrapper[4847]: I1210 14:30:37.110909 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerStarted","Data":"af10217921422cc7ea45c3ab722300a5a8278e5a3fe138f2828840eb7ec03804"} Dec 10 14:30:37 crc kubenswrapper[4847]: I1210 14:30:37.112732 4847 generic.go:334] "Generic (PLEG): container finished" podID="0d91abd9-2ff0-4696-8942-32149e698707" containerID="87f1d6aea560eba1ef453be09d695c66b811d04d13f70693cfa32579bf0926cf" exitCode=0 Dec 10 14:30:37 crc kubenswrapper[4847]: I1210 14:30:37.112765 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerDied","Data":"87f1d6aea560eba1ef453be09d695c66b811d04d13f70693cfa32579bf0926cf"} Dec 10 14:30:37 crc kubenswrapper[4847]: I1210 14:30:37.147355 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r78kt" podStartSLOduration=2.644867086 podStartE2EDuration="4.147332395s" podCreationTimestamp="2025-12-10 14:30:33 +0000 UTC" firstStartedPulling="2025-12-10 14:30:35.096325876 +0000 UTC m=+384.665543506" lastFinishedPulling="2025-12-10 14:30:36.598791185 +0000 UTC m=+386.168008815" observedRunningTime="2025-12-10 14:30:37.134439067 +0000 UTC m=+386.703656717" watchObservedRunningTime="2025-12-10 14:30:37.147332395 +0000 UTC m=+386.716550025" Dec 10 14:30:39 crc kubenswrapper[4847]: I1210 14:30:39.126991 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerStarted","Data":"ac07a3f731a912f09e8666e9cba3737e28d16719a0a6813fce4536c2dd7402f6"} Dec 10 14:30:39 crc kubenswrapper[4847]: I1210 14:30:39.128998 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerStarted","Data":"2fea80235bef64c4646a08eb3f390fe08efbd938575e71c1d2b19878cf27f51d"} Dec 10 14:30:39 crc kubenswrapper[4847]: I1210 14:30:39.152853 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tc2pm" podStartSLOduration=3.150777867 podStartE2EDuration="8.152833393s" podCreationTimestamp="2025-12-10 14:30:31 +0000 UTC" firstStartedPulling="2025-12-10 14:30:33.064743425 +0000 UTC m=+382.633961055" lastFinishedPulling="2025-12-10 14:30:38.066798951 +0000 UTC m=+387.636016581" observedRunningTime="2025-12-10 14:30:39.152195425 +0000 UTC m=+388.721413065" watchObservedRunningTime="2025-12-10 14:30:39.152833393 +0000 UTC m=+388.722051023" Dec 10 14:30:39 crc kubenswrapper[4847]: I1210 14:30:39.170216 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s6wpx" podStartSLOduration=2.949127335 podStartE2EDuration="6.170197124s" podCreationTimestamp="2025-12-10 14:30:33 +0000 UTC" firstStartedPulling="2025-12-10 14:30:35.089323431 +0000 UTC m=+384.658541061" lastFinishedPulling="2025-12-10 14:30:38.31039322 +0000 UTC m=+387.879610850" observedRunningTime="2025-12-10 14:30:39.168498826 +0000 UTC m=+388.737716466" watchObservedRunningTime="2025-12-10 14:30:39.170197124 +0000 UTC m=+388.739414754" Dec 10 14:30:41 crc kubenswrapper[4847]: I1210 14:30:41.528914 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:41 crc kubenswrapper[4847]: I1210 14:30:41.528982 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:41 crc kubenswrapper[4847]: I1210 14:30:41.572479 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:41 crc kubenswrapper[4847]: I1210 14:30:41.723728 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:41 crc kubenswrapper[4847]: I1210 14:30:41.723793 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:42 crc kubenswrapper[4847]: I1210 14:30:42.183155 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4frbk" Dec 10 14:30:42 crc kubenswrapper[4847]: I1210 14:30:42.758563 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tc2pm" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="registry-server" probeResult="failure" output=< Dec 10 14:30:42 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:30:42 crc kubenswrapper[4847]: > Dec 10 14:30:43 crc kubenswrapper[4847]: I1210 14:30:43.932568 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:43 crc kubenswrapper[4847]: I1210 14:30:43.933139 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:43 crc kubenswrapper[4847]: I1210 14:30:43.972365 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:44 crc kubenswrapper[4847]: I1210 14:30:44.129909 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:44 crc kubenswrapper[4847]: I1210 14:30:44.129993 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:44 crc kubenswrapper[4847]: I1210 14:30:44.165499 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:44 crc kubenswrapper[4847]: I1210 14:30:44.192258 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 14:30:44 crc kubenswrapper[4847]: I1210 14:30:44.220570 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:30:51 crc kubenswrapper[4847]: I1210 14:30:51.773485 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:51 crc kubenswrapper[4847]: I1210 14:30:51.819203 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:30:55 crc kubenswrapper[4847]: I1210 14:30:55.081525 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" podUID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" containerName="registry" containerID="cri-o://81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205" gracePeriod=30 Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.049899 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.165974 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.166108 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.166147 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.166220 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.166249 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.166271 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167269 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z4mq\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167443 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\" (UID: \"7ef1bd97-be64-4bee-ae8c-06fd651ac197\") " Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167459 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167498 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167751 4847 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.167775 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ef1bd97-be64-4bee-ae8c-06fd651ac197-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.175178 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.175224 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.175353 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.178224 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq" (OuterVolumeSpecName: "kube-api-access-7z4mq") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "kube-api-access-7z4mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.180623 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.186099 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7ef1bd97-be64-4bee-ae8c-06fd651ac197" (UID: "7ef1bd97-be64-4bee-ae8c-06fd651ac197"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.220345 4847 generic.go:334] "Generic (PLEG): container finished" podID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" containerID="81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205" exitCode=0 Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.220397 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" event={"ID":"7ef1bd97-be64-4bee-ae8c-06fd651ac197","Type":"ContainerDied","Data":"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205"} Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.220412 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.220430 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbhc4" event={"ID":"7ef1bd97-be64-4bee-ae8c-06fd651ac197","Type":"ContainerDied","Data":"6bc09c624c6b917c218b0bfd9eb41240b4039c1a206e9d2e22d48c134b4f8883"} Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.220454 4847 scope.go:117] "RemoveContainer" containerID="81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.238650 4847 scope.go:117] "RemoveContainer" containerID="81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205" Dec 10 14:30:56 crc kubenswrapper[4847]: E1210 14:30:56.239694 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205\": container with ID starting with 81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205 not found: ID does not exist" containerID="81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.239757 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205"} err="failed to get container status \"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205\": rpc error: code = NotFound desc = could not find container \"81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205\": container with ID starting with 81f554a6bf090039455eaf93a3050542b73abd5177d1c3c526fe8b6e0b660205 not found: ID does not exist" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.252493 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.256129 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbhc4"] Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.268670 4847 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7ef1bd97-be64-4bee-ae8c-06fd651ac197-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.268705 4847 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7ef1bd97-be64-4bee-ae8c-06fd651ac197-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.268765 4847 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.268777 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z4mq\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-kube-api-access-7z4mq\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.268786 4847 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7ef1bd97-be64-4bee-ae8c-06fd651ac197-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:30:56 crc kubenswrapper[4847]: I1210 14:30:56.767075 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" path="/var/lib/kubelet/pods/7ef1bd97-be64-4bee-ae8c-06fd651ac197/volumes" Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.010703 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.010768 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.010808 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.011372 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.011425 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a" gracePeriod=600 Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.247119 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a" exitCode=0 Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.247158 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a"} Dec 10 14:31:01 crc kubenswrapper[4847]: I1210 14:31:01.247512 4847 scope.go:117] "RemoveContainer" containerID="6ff55959b80452e45c28dd8fd789764cb6b87f64500e8fd3c618e7164454e6a7" Dec 10 14:31:02 crc kubenswrapper[4847]: I1210 14:31:02.254893 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d"} Dec 10 14:33:01 crc kubenswrapper[4847]: I1210 14:33:01.012239 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:33:01 crc kubenswrapper[4847]: I1210 14:33:01.013197 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:33:31 crc kubenswrapper[4847]: I1210 14:33:31.011163 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:33:31 crc kubenswrapper[4847]: I1210 14:33:31.011780 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.011139 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.011748 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.011810 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.012480 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.012549 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d" gracePeriod=600 Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.876371 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d" exitCode=0 Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.876487 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d"} Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.876762 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5"} Dec 10 14:34:01 crc kubenswrapper[4847]: I1210 14:34:01.876788 4847 scope.go:117] "RemoveContainer" containerID="e455cad56e3de9aa1fe59691717af04d515d603ce2bae757c31491a503cb014a" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.005394 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g8v52"] Dec 10 14:35:36 crc kubenswrapper[4847]: E1210 14:35:36.006156 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" containerName="registry" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.006170 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" containerName="registry" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.006250 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef1bd97-be64-4bee-ae8c-06fd651ac197" containerName="registry" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.006605 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.009381 4847 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8btn6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.009746 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.010665 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.018605 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6nfcw"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.019250 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6nfcw" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.023437 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g8v52"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.023994 4847 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vgffj" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.033288 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7h9v6"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.034194 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.036462 4847 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2bn4z" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.042728 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6nfcw"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.062141 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7h9v6"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.115370 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvrc7\" (UniqueName: \"kubernetes.io/projected/ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d-kube-api-access-mvrc7\") pod \"cert-manager-cainjector-7f985d654d-g8v52\" (UID: \"ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.217535 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dncw\" (UniqueName: \"kubernetes.io/projected/2266f4a5-b282-4686-9a2c-e51d538a11cd-kube-api-access-5dncw\") pod \"cert-manager-webhook-5655c58dd6-7h9v6\" (UID: \"2266f4a5-b282-4686-9a2c-e51d538a11cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.217640 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8prf\" (UniqueName: \"kubernetes.io/projected/f0cacf0f-d6c7-45c1-9241-7a3eb577594c-kube-api-access-s8prf\") pod \"cert-manager-5b446d88c5-6nfcw\" (UID: \"f0cacf0f-d6c7-45c1-9241-7a3eb577594c\") " pod="cert-manager/cert-manager-5b446d88c5-6nfcw" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.217707 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvrc7\" (UniqueName: \"kubernetes.io/projected/ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d-kube-api-access-mvrc7\") pod \"cert-manager-cainjector-7f985d654d-g8v52\" (UID: \"ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.249431 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvrc7\" (UniqueName: \"kubernetes.io/projected/ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d-kube-api-access-mvrc7\") pod \"cert-manager-cainjector-7f985d654d-g8v52\" (UID: \"ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.318769 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8prf\" (UniqueName: \"kubernetes.io/projected/f0cacf0f-d6c7-45c1-9241-7a3eb577594c-kube-api-access-s8prf\") pod \"cert-manager-5b446d88c5-6nfcw\" (UID: \"f0cacf0f-d6c7-45c1-9241-7a3eb577594c\") " pod="cert-manager/cert-manager-5b446d88c5-6nfcw" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.318915 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dncw\" (UniqueName: \"kubernetes.io/projected/2266f4a5-b282-4686-9a2c-e51d538a11cd-kube-api-access-5dncw\") pod \"cert-manager-webhook-5655c58dd6-7h9v6\" (UID: \"2266f4a5-b282-4686-9a2c-e51d538a11cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.323118 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.337560 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8prf\" (UniqueName: \"kubernetes.io/projected/f0cacf0f-d6c7-45c1-9241-7a3eb577594c-kube-api-access-s8prf\") pod \"cert-manager-5b446d88c5-6nfcw\" (UID: \"f0cacf0f-d6c7-45c1-9241-7a3eb577594c\") " pod="cert-manager/cert-manager-5b446d88c5-6nfcw" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.337890 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dncw\" (UniqueName: \"kubernetes.io/projected/2266f4a5-b282-4686-9a2c-e51d538a11cd-kube-api-access-5dncw\") pod \"cert-manager-webhook-5655c58dd6-7h9v6\" (UID: \"2266f4a5-b282-4686-9a2c-e51d538a11cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.347456 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.551427 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-g8v52"] Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.585056 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.636039 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6nfcw" Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.837576 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7h9v6"] Dec 10 14:35:36 crc kubenswrapper[4847]: W1210 14:35:36.838562 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2266f4a5_b282_4686_9a2c_e51d538a11cd.slice/crio-591febe64284a9158f07737a5ad19811665c9011de4fc50bea0ecb5beca9844f WatchSource:0}: Error finding container 591febe64284a9158f07737a5ad19811665c9011de4fc50bea0ecb5beca9844f: Status 404 returned error can't find the container with id 591febe64284a9158f07737a5ad19811665c9011de4fc50bea0ecb5beca9844f Dec 10 14:35:36 crc kubenswrapper[4847]: I1210 14:35:36.845814 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6nfcw"] Dec 10 14:35:36 crc kubenswrapper[4847]: W1210 14:35:36.852922 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0cacf0f_d6c7_45c1_9241_7a3eb577594c.slice/crio-b69ebb00ae4404b5cf27f9f509778f4e5235e77fc34db962dc833ae46b019f62 WatchSource:0}: Error finding container b69ebb00ae4404b5cf27f9f509778f4e5235e77fc34db962dc833ae46b019f62: Status 404 returned error can't find the container with id b69ebb00ae4404b5cf27f9f509778f4e5235e77fc34db962dc833ae46b019f62 Dec 10 14:35:37 crc kubenswrapper[4847]: I1210 14:35:37.460595 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6nfcw" event={"ID":"f0cacf0f-d6c7-45c1-9241-7a3eb577594c","Type":"ContainerStarted","Data":"b69ebb00ae4404b5cf27f9f509778f4e5235e77fc34db962dc833ae46b019f62"} Dec 10 14:35:37 crc kubenswrapper[4847]: I1210 14:35:37.462531 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" event={"ID":"ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d","Type":"ContainerStarted","Data":"8d90be1754f5bc2896f62bbfdadc6f151e048c45524afe251a67c503edfed886"} Dec 10 14:35:37 crc kubenswrapper[4847]: I1210 14:35:37.462634 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" event={"ID":"2266f4a5-b282-4686-9a2c-e51d538a11cd","Type":"ContainerStarted","Data":"591febe64284a9158f07737a5ad19811665c9011de4fc50bea0ecb5beca9844f"} Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.489914 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" event={"ID":"ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d","Type":"ContainerStarted","Data":"5fa1b4289976c3d92e517ccc837d499513afead2f0e5542de79cd0021164b444"} Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.491080 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" event={"ID":"2266f4a5-b282-4686-9a2c-e51d538a11cd","Type":"ContainerStarted","Data":"01c2bbfd85af5742fa118e5ecea7f497dc26f8396dede41328f2aa40bd79119c"} Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.491851 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.493088 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6nfcw" event={"ID":"f0cacf0f-d6c7-45c1-9241-7a3eb577594c","Type":"ContainerStarted","Data":"ecad6533416f7bc8eeb54b0e190c822fcba5057fc08b885eb7fe93c14e0580dc"} Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.503891 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-g8v52" podStartSLOduration=2.321072486 podStartE2EDuration="6.503875799s" podCreationTimestamp="2025-12-10 14:35:35 +0000 UTC" firstStartedPulling="2025-12-10 14:35:36.584795768 +0000 UTC m=+686.154013398" lastFinishedPulling="2025-12-10 14:35:40.767599081 +0000 UTC m=+690.336816711" observedRunningTime="2025-12-10 14:35:41.502398298 +0000 UTC m=+691.071615928" watchObservedRunningTime="2025-12-10 14:35:41.503875799 +0000 UTC m=+691.073093429" Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.522887 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" podStartSLOduration=1.5944942800000002 podStartE2EDuration="5.522868299s" podCreationTimestamp="2025-12-10 14:35:36 +0000 UTC" firstStartedPulling="2025-12-10 14:35:36.842106032 +0000 UTC m=+686.411323662" lastFinishedPulling="2025-12-10 14:35:40.770480051 +0000 UTC m=+690.339697681" observedRunningTime="2025-12-10 14:35:41.520478202 +0000 UTC m=+691.089695842" watchObservedRunningTime="2025-12-10 14:35:41.522868299 +0000 UTC m=+691.092085929" Dec 10 14:35:41 crc kubenswrapper[4847]: I1210 14:35:41.555023 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-6nfcw" podStartSLOduration=2.55071207 podStartE2EDuration="6.554991112s" podCreationTimestamp="2025-12-10 14:35:35 +0000 UTC" firstStartedPulling="2025-12-10 14:35:36.855654199 +0000 UTC m=+686.424871829" lastFinishedPulling="2025-12-10 14:35:40.859933241 +0000 UTC m=+690.429150871" observedRunningTime="2025-12-10 14:35:41.547742311 +0000 UTC m=+691.116959951" watchObservedRunningTime="2025-12-10 14:35:41.554991112 +0000 UTC m=+691.124208752" Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.356993 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-7h9v6" Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.790431 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b5l5q"] Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791067 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-controller" containerID="cri-o://1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791169 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="nbdb" containerID="cri-o://682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791229 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791307 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-node" containerID="cri-o://6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791213 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="northd" containerID="cri-o://a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791356 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-acl-logging" containerID="cri-o://b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.791437 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="sbdb" containerID="cri-o://8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" gracePeriod=30 Dec 10 14:35:46 crc kubenswrapper[4847]: I1210 14:35:46.845803 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" containerID="cri-o://c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" gracePeriod=30 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.123617 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/3.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.126561 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovn-acl-logging/0.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.127283 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovn-controller/0.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.127789 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184352 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9tsqg"] Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184564 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184576 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184584 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="northd" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184592 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="northd" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184600 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184606 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184616 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184623 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184630 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-acl-logging" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184638 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-acl-logging" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184648 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184654 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184667 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184674 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184685 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="sbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184693 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="sbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184707 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="nbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184734 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="nbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184743 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184749 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184758 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kubecfg-setup" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184764 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kubecfg-setup" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.184773 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-node" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184779 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-node" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184867 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184879 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184887 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-acl-logging" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184900 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-node" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184908 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="sbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184915 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184927 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="nbdb" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184935 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="northd" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184945 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovn-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.184956 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.185043 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.185049 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.185131 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.185141 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerName="ovnkube-controller" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.186879 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.269904 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.269995 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270036 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270060 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270099 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270163 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270251 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270301 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270329 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270357 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270377 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270421 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270427 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log" (OuterVolumeSpecName: "node-log") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270469 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270495 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270491 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket" (OuterVolumeSpecName: "log-socket") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270473 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270532 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270491 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270572 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270536 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270574 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270604 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270667 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270698 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270701 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270740 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash" (OuterVolumeSpecName: "host-slash") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270822 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270801 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270743 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270882 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270903 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhhpv\" (UniqueName: \"kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270955 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270976 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert\") pod \"9b405b3d-ae75-409a-9e83-b098f333a5c0\" (UID: \"9b405b3d-ae75-409a-9e83-b098f333a5c0\") " Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.270976 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271164 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-var-lib-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271192 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-env-overrides\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271232 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-bin\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271265 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-kubelet\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271279 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-slash\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271297 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhvvj\" (UniqueName: \"kubernetes.io/projected/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-kube-api-access-lhvvj\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271327 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-netd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271360 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-systemd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271393 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovn-node-metrics-cert\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271420 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-ovn\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271434 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-log-socket\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271450 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-config\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271448 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271519 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271599 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-node-log\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271652 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271694 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-netns\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271794 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271839 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-systemd-units\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271888 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-etc-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271921 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.271962 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-script-lib\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272065 4847 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272098 4847 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272128 4847 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272152 4847 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272170 4847 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272189 4847 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-slash\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272207 4847 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272224 4847 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272242 4847 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272260 4847 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272278 4847 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272298 4847 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272316 4847 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9b405b3d-ae75-409a-9e83-b098f333a5c0-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272334 4847 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-node-log\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272355 4847 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272375 4847 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.272393 4847 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-log-socket\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.276983 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.277618 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv" (OuterVolumeSpecName: "kube-api-access-jhhpv") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "kube-api-access-jhhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.287607 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9b405b3d-ae75-409a-9e83-b098f333a5c0" (UID: "9b405b3d-ae75-409a-9e83-b098f333a5c0"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373313 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-etc-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373379 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373405 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-script-lib\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373439 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-var-lib-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373464 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-env-overrides\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373495 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-bin\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373498 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-etc-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373563 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-kubelet\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373571 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-var-lib-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373522 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-kubelet\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373632 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373620 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-bin\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373705 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-slash\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhvvj\" (UniqueName: \"kubernetes.io/projected/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-kube-api-access-lhvvj\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373855 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-slash\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373933 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-netd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.373980 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-cni-netd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374013 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-systemd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374071 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-systemd\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374094 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovn-node-metrics-cert\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374138 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-ovn\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374170 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-log-socket\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374196 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-config\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374225 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-env-overrides\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374264 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-node-log\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374294 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-log-socket\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374297 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-node-log\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374289 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-script-lib\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374318 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374299 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-ovn\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374405 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-netns\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374425 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-netns\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374427 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-run-openvswitch\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374442 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374479 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-host-run-ovn-kubernetes\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374507 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-systemd-units\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374582 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-systemd-units\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374606 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhhpv\" (UniqueName: \"kubernetes.io/projected/9b405b3d-ae75-409a-9e83-b098f333a5c0-kube-api-access-jhhpv\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374624 4847 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9b405b3d-ae75-409a-9e83-b098f333a5c0-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374640 4847 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9b405b3d-ae75-409a-9e83-b098f333a5c0-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.374893 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovnkube-config\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.377278 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-ovn-node-metrics-cert\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.390664 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhvvj\" (UniqueName: \"kubernetes.io/projected/0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564-kube-api-access-lhvvj\") pod \"ovnkube-node-9tsqg\" (UID: \"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564\") " pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.501925 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.531540 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovnkube-controller/3.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.533974 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovn-acl-logging/0.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535006 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-b5l5q_9b405b3d-ae75-409a-9e83-b098f333a5c0/ovn-controller/0.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535320 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535352 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535362 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535360 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535410 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535423 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535433 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535451 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535374 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535477 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535488 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" exitCode=0 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535497 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" exitCode=143 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535504 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b405b3d-ae75-409a-9e83-b098f333a5c0" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" exitCode=143 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535504 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535544 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535561 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535582 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535592 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535600 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535510 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535607 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535747 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535767 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535775 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535783 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535805 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535831 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535838 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535843 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535849 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535855 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535861 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535866 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535871 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535877 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535882 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535890 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535900 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535909 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535915 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535920 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535926 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535931 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535936 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535942 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535946 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535952 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535962 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b5l5q" event={"ID":"9b405b3d-ae75-409a-9e83-b098f333a5c0","Type":"ContainerDied","Data":"a1efc85f1dd13bc6a32c659829e3afc2b08af6502658d64d2236aa23de9dd8fd"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535973 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535978 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535983 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535988 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535993 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.535998 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.536004 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.536009 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.536015 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.536020 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.537931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"31ad7e6cc02ff73114bd70e4f66b44dc688fcb3934a1266b78b62e0cba9ebaf4"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.542877 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/2.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.543789 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/1.log" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.543832 4847 generic.go:334] "Generic (PLEG): container finished" podID="71681159-7da8-4bc9-837c-d0e3b7397e2e" containerID="a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a" exitCode=2 Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.543861 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerDied","Data":"a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.543885 4847 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67"} Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.544833 4847 scope.go:117] "RemoveContainer" containerID="a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.545147 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-67b5l_openshift-multus(71681159-7da8-4bc9-837c-d0e3b7397e2e)\"" pod="openshift-multus/multus-67b5l" podUID="71681159-7da8-4bc9-837c-d0e3b7397e2e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.577501 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.591592 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b5l5q"] Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.607753 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b5l5q"] Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.609927 4847 scope.go:117] "RemoveContainer" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.625951 4847 scope.go:117] "RemoveContainer" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.645828 4847 scope.go:117] "RemoveContainer" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.661347 4847 scope.go:117] "RemoveContainer" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.677347 4847 scope.go:117] "RemoveContainer" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.759467 4847 scope.go:117] "RemoveContainer" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.783165 4847 scope.go:117] "RemoveContainer" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.803242 4847 scope.go:117] "RemoveContainer" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.820667 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.821174 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.821261 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} err="failed to get container status \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.821329 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.821937 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": container with ID starting with e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51 not found: ID does not exist" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.821992 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} err="failed to get container status \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": rpc error: code = NotFound desc = could not find container \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": container with ID starting with e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.822036 4847 scope.go:117] "RemoveContainer" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.822417 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": container with ID starting with 8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7 not found: ID does not exist" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.822450 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} err="failed to get container status \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": rpc error: code = NotFound desc = could not find container \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": container with ID starting with 8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.822474 4847 scope.go:117] "RemoveContainer" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.822907 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": container with ID starting with 682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886 not found: ID does not exist" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.822940 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} err="failed to get container status \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": rpc error: code = NotFound desc = could not find container \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": container with ID starting with 682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.822963 4847 scope.go:117] "RemoveContainer" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.823414 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": container with ID starting with a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d not found: ID does not exist" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.823458 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} err="failed to get container status \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": rpc error: code = NotFound desc = could not find container \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": container with ID starting with a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.823488 4847 scope.go:117] "RemoveContainer" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.823907 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": container with ID starting with 849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7 not found: ID does not exist" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.823946 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} err="failed to get container status \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": rpc error: code = NotFound desc = could not find container \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": container with ID starting with 849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.823969 4847 scope.go:117] "RemoveContainer" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.824353 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": container with ID starting with 6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e not found: ID does not exist" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.824385 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} err="failed to get container status \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": rpc error: code = NotFound desc = could not find container \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": container with ID starting with 6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.824404 4847 scope.go:117] "RemoveContainer" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.824934 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": container with ID starting with b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7 not found: ID does not exist" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.824956 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} err="failed to get container status \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": rpc error: code = NotFound desc = could not find container \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": container with ID starting with b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.824970 4847 scope.go:117] "RemoveContainer" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.825381 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": container with ID starting with 1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7 not found: ID does not exist" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.825432 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} err="failed to get container status \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": rpc error: code = NotFound desc = could not find container \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": container with ID starting with 1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.825467 4847 scope.go:117] "RemoveContainer" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: E1210 14:35:47.825992 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": container with ID starting with aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0 not found: ID does not exist" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.826021 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} err="failed to get container status \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": rpc error: code = NotFound desc = could not find container \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": container with ID starting with aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.826038 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.826598 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} err="failed to get container status \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.826640 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.827220 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} err="failed to get container status \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": rpc error: code = NotFound desc = could not find container \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": container with ID starting with e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.827244 4847 scope.go:117] "RemoveContainer" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.827661 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} err="failed to get container status \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": rpc error: code = NotFound desc = could not find container \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": container with ID starting with 8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.827702 4847 scope.go:117] "RemoveContainer" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.828120 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} err="failed to get container status \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": rpc error: code = NotFound desc = could not find container \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": container with ID starting with 682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.828152 4847 scope.go:117] "RemoveContainer" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.828541 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} err="failed to get container status \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": rpc error: code = NotFound desc = could not find container \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": container with ID starting with a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.828569 4847 scope.go:117] "RemoveContainer" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.829064 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} err="failed to get container status \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": rpc error: code = NotFound desc = could not find container \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": container with ID starting with 849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.829103 4847 scope.go:117] "RemoveContainer" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.829439 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} err="failed to get container status \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": rpc error: code = NotFound desc = could not find container \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": container with ID starting with 6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.829463 4847 scope.go:117] "RemoveContainer" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.830181 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} err="failed to get container status \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": rpc error: code = NotFound desc = could not find container \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": container with ID starting with b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.830210 4847 scope.go:117] "RemoveContainer" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.830691 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} err="failed to get container status \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": rpc error: code = NotFound desc = could not find container \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": container with ID starting with 1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.830767 4847 scope.go:117] "RemoveContainer" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.831265 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} err="failed to get container status \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": rpc error: code = NotFound desc = could not find container \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": container with ID starting with aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.831301 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.831681 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} err="failed to get container status \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.831734 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.832234 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} err="failed to get container status \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": rpc error: code = NotFound desc = could not find container \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": container with ID starting with e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.832258 4847 scope.go:117] "RemoveContainer" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.832623 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} err="failed to get container status \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": rpc error: code = NotFound desc = could not find container \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": container with ID starting with 8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.832644 4847 scope.go:117] "RemoveContainer" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833209 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} err="failed to get container status \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": rpc error: code = NotFound desc = could not find container \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": container with ID starting with 682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833240 4847 scope.go:117] "RemoveContainer" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833562 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} err="failed to get container status \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": rpc error: code = NotFound desc = could not find container \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": container with ID starting with a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833596 4847 scope.go:117] "RemoveContainer" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833941 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} err="failed to get container status \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": rpc error: code = NotFound desc = could not find container \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": container with ID starting with 849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.833972 4847 scope.go:117] "RemoveContainer" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.834236 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} err="failed to get container status \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": rpc error: code = NotFound desc = could not find container \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": container with ID starting with 6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.834257 4847 scope.go:117] "RemoveContainer" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.834579 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} err="failed to get container status \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": rpc error: code = NotFound desc = could not find container \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": container with ID starting with b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.834597 4847 scope.go:117] "RemoveContainer" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.834910 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} err="failed to get container status \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": rpc error: code = NotFound desc = could not find container \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": container with ID starting with 1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835004 4847 scope.go:117] "RemoveContainer" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835291 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} err="failed to get container status \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": rpc error: code = NotFound desc = could not find container \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": container with ID starting with aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835313 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835529 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} err="failed to get container status \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835545 4847 scope.go:117] "RemoveContainer" containerID="e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835811 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51"} err="failed to get container status \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": rpc error: code = NotFound desc = could not find container \"e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51\": container with ID starting with e7db2de5b5c83b7d727242d418b2df324092f58291b4371c19c0eca39693bc51 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.835848 4847 scope.go:117] "RemoveContainer" containerID="8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.836725 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7"} err="failed to get container status \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": rpc error: code = NotFound desc = could not find container \"8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7\": container with ID starting with 8031504b6ee3de9c704dbd0db0d3791254890e14352dc344bfaac0bf0e4a65b7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.836791 4847 scope.go:117] "RemoveContainer" containerID="682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837140 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886"} err="failed to get container status \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": rpc error: code = NotFound desc = could not find container \"682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886\": container with ID starting with 682fcf3a807bcd23fa513598c00d9bf34626767a09dfd07bf0a6a12066c14886 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837161 4847 scope.go:117] "RemoveContainer" containerID="a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837394 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d"} err="failed to get container status \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": rpc error: code = NotFound desc = could not find container \"a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d\": container with ID starting with a5ec075d640f28adf72ed13e86a9954aec4c8cef065322809a497bc8d766354d not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837417 4847 scope.go:117] "RemoveContainer" containerID="849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837629 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7"} err="failed to get container status \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": rpc error: code = NotFound desc = could not find container \"849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7\": container with ID starting with 849f85425fefead5a883777d3ce61cd7778cea5e83f624652c0fd89c63e509d7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.837647 4847 scope.go:117] "RemoveContainer" containerID="6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838024 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e"} err="failed to get container status \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": rpc error: code = NotFound desc = could not find container \"6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e\": container with ID starting with 6fdbd2f15606bc1f85e1464236c4bab99af6fe82efb56f147eef3c7cc058060e not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838044 4847 scope.go:117] "RemoveContainer" containerID="b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838274 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7"} err="failed to get container status \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": rpc error: code = NotFound desc = could not find container \"b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7\": container with ID starting with b3d1e57d07f1f09e6c99fd194b432fbbcc49cefc7b2a41f0ad4c8ff8bbf8c9c7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838290 4847 scope.go:117] "RemoveContainer" containerID="1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838585 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7"} err="failed to get container status \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": rpc error: code = NotFound desc = could not find container \"1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7\": container with ID starting with 1d9ca810ca2eee1797eddf169f12c7d7b8899e971bf66b8584302bf8894caea7 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838602 4847 scope.go:117] "RemoveContainer" containerID="aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838860 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0"} err="failed to get container status \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": rpc error: code = NotFound desc = could not find container \"aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0\": container with ID starting with aee6407f1f23d14773b79ca9c213dfdb40e26bffab018d0e84fdd911f9e9bad0 not found: ID does not exist" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.838876 4847 scope.go:117] "RemoveContainer" containerID="c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6" Dec 10 14:35:47 crc kubenswrapper[4847]: I1210 14:35:47.839120 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6"} err="failed to get container status \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": rpc error: code = NotFound desc = could not find container \"c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6\": container with ID starting with c216f9563e504be1427f3c2222ca1305ff2e9da6a25627e1a43b55570011cac6 not found: ID does not exist" Dec 10 14:35:48 crc kubenswrapper[4847]: I1210 14:35:48.552245 4847 generic.go:334] "Generic (PLEG): container finished" podID="0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564" containerID="b2b802af2e6dce7b7d0a95ef50fc27039e47bc18a22c340bfe895742bfbcb0a8" exitCode=0 Dec 10 14:35:48 crc kubenswrapper[4847]: I1210 14:35:48.552352 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerDied","Data":"b2b802af2e6dce7b7d0a95ef50fc27039e47bc18a22c340bfe895742bfbcb0a8"} Dec 10 14:35:48 crc kubenswrapper[4847]: I1210 14:35:48.765854 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b405b3d-ae75-409a-9e83-b098f333a5c0" path="/var/lib/kubelet/pods/9b405b3d-ae75-409a-9e83-b098f333a5c0/volumes" Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.562938 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"7c226b2bf527559efc1da6e1ebe0762eb67445578efda3a3224c920e4a8e616d"} Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.563312 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"e20ea1029699a148e7cdbaad3cfe5051e0c202d03619778c9d8d87e749bfbb2b"} Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.563325 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"e819f705bb2442ac718f08a0cc7929a90ca4db87f317b7ecc040022ba75b1602"} Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.563336 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"69a8a3496c2c07d94a675d5b6db5d6c209195a991063d4d53f4237842b7b0419"} Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.563345 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"a7540831edef36da5fe9c3019976bd65b42c25ca58b61a445fa1bbe9f06c8886"} Dec 10 14:35:49 crc kubenswrapper[4847]: I1210 14:35:49.563355 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"65c9e5fb95181a547db786163589d407dca7a0a90cba603567a6d7fc594e87fa"} Dec 10 14:35:51 crc kubenswrapper[4847]: I1210 14:35:51.576307 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"d04c7f064825ddc835c0c6a979276646246789bc07765f650939be28ff319385"} Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.591300 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" event={"ID":"0dc4f6f9-e7ee-4ac1-b1c6-3163ef968564","Type":"ContainerStarted","Data":"914e7d7d2a8763f031df917ee02e7954fdabc47f0b7a021a8079a5f0352bd2f7"} Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.591932 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.591954 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.591965 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.623081 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.623166 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:35:53 crc kubenswrapper[4847]: I1210 14:35:53.639369 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" podStartSLOduration=6.639333222 podStartE2EDuration="6.639333222s" podCreationTimestamp="2025-12-10 14:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:35:53.635022542 +0000 UTC m=+703.204240192" watchObservedRunningTime="2025-12-10 14:35:53.639333222 +0000 UTC m=+703.208550852" Dec 10 14:35:59 crc kubenswrapper[4847]: I1210 14:35:59.759890 4847 scope.go:117] "RemoveContainer" containerID="a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a" Dec 10 14:35:59 crc kubenswrapper[4847]: E1210 14:35:59.760681 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-67b5l_openshift-multus(71681159-7da8-4bc9-837c-d0e3b7397e2e)\"" pod="openshift-multus/multus-67b5l" podUID="71681159-7da8-4bc9-837c-d0e3b7397e2e" Dec 10 14:36:01 crc kubenswrapper[4847]: I1210 14:36:01.012010 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:36:01 crc kubenswrapper[4847]: I1210 14:36:01.012941 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:36:11 crc kubenswrapper[4847]: I1210 14:36:11.063074 4847 scope.go:117] "RemoveContainer" containerID="c591ec5832530c8df60976544dbf475ef1000886482b841307a786f8173fef67" Dec 10 14:36:11 crc kubenswrapper[4847]: I1210 14:36:11.706633 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/2.log" Dec 10 14:36:13 crc kubenswrapper[4847]: I1210 14:36:13.760247 4847 scope.go:117] "RemoveContainer" containerID="a62839b930ae21957a8c42388e0dc68fae3d572d27b54c8325611a0ae601fa7a" Dec 10 14:36:14 crc kubenswrapper[4847]: I1210 14:36:14.729045 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-67b5l_71681159-7da8-4bc9-837c-d0e3b7397e2e/kube-multus/2.log" Dec 10 14:36:14 crc kubenswrapper[4847]: I1210 14:36:14.729784 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-67b5l" event={"ID":"71681159-7da8-4bc9-837c-d0e3b7397e2e","Type":"ContainerStarted","Data":"a28978cf689c811a239d2471c4816bb25e1af4f0d63616a6b93b23f6838249d9"} Dec 10 14:36:17 crc kubenswrapper[4847]: I1210 14:36:17.530129 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9tsqg" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.148087 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz"] Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.150015 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.161185 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz"] Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.165328 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.257781 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzd49\" (UniqueName: \"kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.257888 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.258171 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.359661 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.359764 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzd49\" (UniqueName: \"kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.359793 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.360418 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.360464 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.382883 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzd49\" (UniqueName: \"kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.482124 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.689044 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz"] Dec 10 14:36:24 crc kubenswrapper[4847]: I1210 14:36:24.781033 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" event={"ID":"b5094984-5419-4b07-bbfd-033eb535c588","Type":"ContainerStarted","Data":"4416c2879b706305c834ad47d7641ac795ccc44402eae4779d89a448309884e7"} Dec 10 14:36:26 crc kubenswrapper[4847]: I1210 14:36:26.793233 4847 generic.go:334] "Generic (PLEG): container finished" podID="b5094984-5419-4b07-bbfd-033eb535c588" containerID="b749ce7130cd0a40989306c0fcbd9d78a2500269097bfc95fc5892b3e446e28f" exitCode=0 Dec 10 14:36:26 crc kubenswrapper[4847]: I1210 14:36:26.793381 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" event={"ID":"b5094984-5419-4b07-bbfd-033eb535c588","Type":"ContainerDied","Data":"b749ce7130cd0a40989306c0fcbd9d78a2500269097bfc95fc5892b3e446e28f"} Dec 10 14:36:28 crc kubenswrapper[4847]: I1210 14:36:28.807347 4847 generic.go:334] "Generic (PLEG): container finished" podID="b5094984-5419-4b07-bbfd-033eb535c588" containerID="4252279ccd5567fb31c5654436498af7e34b568ffbe1f769cd01e0bc2238ff93" exitCode=0 Dec 10 14:36:28 crc kubenswrapper[4847]: I1210 14:36:28.807404 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" event={"ID":"b5094984-5419-4b07-bbfd-033eb535c588","Type":"ContainerDied","Data":"4252279ccd5567fb31c5654436498af7e34b568ffbe1f769cd01e0bc2238ff93"} Dec 10 14:36:29 crc kubenswrapper[4847]: I1210 14:36:29.816295 4847 generic.go:334] "Generic (PLEG): container finished" podID="b5094984-5419-4b07-bbfd-033eb535c588" containerID="71f0e36607bb6381cdc5c480bb0f2ffdc71c8d7d8d9bbde9175fa26b2a410143" exitCode=0 Dec 10 14:36:29 crc kubenswrapper[4847]: I1210 14:36:29.816353 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" event={"ID":"b5094984-5419-4b07-bbfd-033eb535c588","Type":"ContainerDied","Data":"71f0e36607bb6381cdc5c480bb0f2ffdc71c8d7d8d9bbde9175fa26b2a410143"} Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.012009 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.012686 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.038806 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.170893 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzd49\" (UniqueName: \"kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49\") pod \"b5094984-5419-4b07-bbfd-033eb535c588\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.170995 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util\") pod \"b5094984-5419-4b07-bbfd-033eb535c588\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.171053 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle\") pod \"b5094984-5419-4b07-bbfd-033eb535c588\" (UID: \"b5094984-5419-4b07-bbfd-033eb535c588\") " Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.171870 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle" (OuterVolumeSpecName: "bundle") pod "b5094984-5419-4b07-bbfd-033eb535c588" (UID: "b5094984-5419-4b07-bbfd-033eb535c588"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.177164 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49" (OuterVolumeSpecName: "kube-api-access-wzd49") pod "b5094984-5419-4b07-bbfd-033eb535c588" (UID: "b5094984-5419-4b07-bbfd-033eb535c588"). InnerVolumeSpecName "kube-api-access-wzd49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.272248 4847 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.272287 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzd49\" (UniqueName: \"kubernetes.io/projected/b5094984-5419-4b07-bbfd-033eb535c588-kube-api-access-wzd49\") on node \"crc\" DevicePath \"\"" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.448379 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util" (OuterVolumeSpecName: "util") pod "b5094984-5419-4b07-bbfd-033eb535c588" (UID: "b5094984-5419-4b07-bbfd-033eb535c588"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.474243 4847 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b5094984-5419-4b07-bbfd-033eb535c588-util\") on node \"crc\" DevicePath \"\"" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.830121 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" event={"ID":"b5094984-5419-4b07-bbfd-033eb535c588","Type":"ContainerDied","Data":"4416c2879b706305c834ad47d7641ac795ccc44402eae4779d89a448309884e7"} Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.830157 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4416c2879b706305c834ad47d7641ac795ccc44402eae4779d89a448309884e7" Dec 10 14:36:31 crc kubenswrapper[4847]: I1210 14:36:31.830214 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.176402 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc"] Dec 10 14:36:33 crc kubenswrapper[4847]: E1210 14:36:33.176887 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="util" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.176899 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="util" Dec 10 14:36:33 crc kubenswrapper[4847]: E1210 14:36:33.176917 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="pull" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.176922 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="pull" Dec 10 14:36:33 crc kubenswrapper[4847]: E1210 14:36:33.176933 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="extract" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.176940 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="extract" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.177019 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5094984-5419-4b07-bbfd-033eb535c588" containerName="extract" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.177403 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.179184 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-kh85g" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.179426 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.180431 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.193176 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc"] Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.196781 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb4hb\" (UniqueName: \"kubernetes.io/projected/64363456-a381-49ba-98f2-b16b9928b65b-kube-api-access-hb4hb\") pod \"nmstate-operator-5b5b58f5c8-x4fhc\" (UID: \"64363456-a381-49ba-98f2-b16b9928b65b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.297931 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb4hb\" (UniqueName: \"kubernetes.io/projected/64363456-a381-49ba-98f2-b16b9928b65b-kube-api-access-hb4hb\") pod \"nmstate-operator-5b5b58f5c8-x4fhc\" (UID: \"64363456-a381-49ba-98f2-b16b9928b65b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.315043 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb4hb\" (UniqueName: \"kubernetes.io/projected/64363456-a381-49ba-98f2-b16b9928b65b-kube-api-access-hb4hb\") pod \"nmstate-operator-5b5b58f5c8-x4fhc\" (UID: \"64363456-a381-49ba-98f2-b16b9928b65b\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.541671 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.751844 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc"] Dec 10 14:36:33 crc kubenswrapper[4847]: W1210 14:36:33.764074 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64363456_a381_49ba_98f2_b16b9928b65b.slice/crio-85f6baa4fd9e162fef85c729a8842a15fc9cf642ee588534e7e8572530fc59c9 WatchSource:0}: Error finding container 85f6baa4fd9e162fef85c729a8842a15fc9cf642ee588534e7e8572530fc59c9: Status 404 returned error can't find the container with id 85f6baa4fd9e162fef85c729a8842a15fc9cf642ee588534e7e8572530fc59c9 Dec 10 14:36:33 crc kubenswrapper[4847]: I1210 14:36:33.848660 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" event={"ID":"64363456-a381-49ba-98f2-b16b9928b65b","Type":"ContainerStarted","Data":"85f6baa4fd9e162fef85c729a8842a15fc9cf642ee588534e7e8572530fc59c9"} Dec 10 14:36:37 crc kubenswrapper[4847]: I1210 14:36:37.869133 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" event={"ID":"64363456-a381-49ba-98f2-b16b9928b65b","Type":"ContainerStarted","Data":"52b5301b6e0f312957ec5656c18bcf2c541a5c03425135aa1b2244bbec271f86"} Dec 10 14:36:37 crc kubenswrapper[4847]: I1210 14:36:37.891283 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-x4fhc" podStartSLOduration=1.722649136 podStartE2EDuration="4.891264017s" podCreationTimestamp="2025-12-10 14:36:33 +0000 UTC" firstStartedPulling="2025-12-10 14:36:33.76640645 +0000 UTC m=+743.335624080" lastFinishedPulling="2025-12-10 14:36:36.935021331 +0000 UTC m=+746.504238961" observedRunningTime="2025-12-10 14:36:37.886307189 +0000 UTC m=+747.455524839" watchObservedRunningTime="2025-12-10 14:36:37.891264017 +0000 UTC m=+747.460481667" Dec 10 14:36:39 crc kubenswrapper[4847]: I1210 14:36:39.079068 4847 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.228218 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.230646 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.232565 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-kfv7r" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.236297 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.237050 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.242043 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.246131 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.260419 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wxpbs"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.261282 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.272732 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.332117 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7ll\" (UniqueName: \"kubernetes.io/projected/e7bb30c3-5f7b-465a-8852-388f8204277b-kube-api-access-zx7ll\") pod \"nmstate-metrics-7f946cbc9-5rj2v\" (UID: \"e7bb30c3-5f7b-465a-8852-388f8204277b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.332387 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-596dv\" (UniqueName: \"kubernetes.io/projected/ef0ffd00-af07-4843-9b34-2cf613252d7c-kube-api-access-596dv\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.332478 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.361156 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.362028 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.364421 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.365447 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-bbmf4" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.367601 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.370985 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.433930 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7ll\" (UniqueName: \"kubernetes.io/projected/e7bb30c3-5f7b-465a-8852-388f8204277b-kube-api-access-zx7ll\") pod \"nmstate-metrics-7f946cbc9-5rj2v\" (UID: \"e7bb30c3-5f7b-465a-8852-388f8204277b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434021 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-dbus-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434058 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxt2g\" (UniqueName: \"kubernetes.io/projected/b51a4da1-a921-414f-9985-044264a489ad-kube-api-access-kxt2g\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434091 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-nmstate-lock\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434122 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-596dv\" (UniqueName: \"kubernetes.io/projected/ef0ffd00-af07-4843-9b34-2cf613252d7c-kube-api-access-596dv\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434147 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-ovs-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.434176 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: E1210 14:36:44.434317 4847 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 10 14:36:44 crc kubenswrapper[4847]: E1210 14:36:44.434368 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair podName:ef0ffd00-af07-4843-9b34-2cf613252d7c nodeName:}" failed. No retries permitted until 2025-12-10 14:36:44.934348872 +0000 UTC m=+754.503566502 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-4kbcp" (UID: "ef0ffd00-af07-4843-9b34-2cf613252d7c") : secret "openshift-nmstate-webhook" not found Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.452319 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-596dv\" (UniqueName: \"kubernetes.io/projected/ef0ffd00-af07-4843-9b34-2cf613252d7c-kube-api-access-596dv\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.454459 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7ll\" (UniqueName: \"kubernetes.io/projected/e7bb30c3-5f7b-465a-8852-388f8204277b-kube-api-access-zx7ll\") pod \"nmstate-metrics-7f946cbc9-5rj2v\" (UID: \"e7bb30c3-5f7b-465a-8852-388f8204277b\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.535773 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5575bcb-0d29-4538-8213-99a007548880-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536185 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-dbus-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536216 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hspdq\" (UniqueName: \"kubernetes.io/projected/d5575bcb-0d29-4538-8213-99a007548880-kube-api-access-hspdq\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536245 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5575bcb-0d29-4538-8213-99a007548880-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536275 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxt2g\" (UniqueName: \"kubernetes.io/projected/b51a4da1-a921-414f-9985-044264a489ad-kube-api-access-kxt2g\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536306 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-nmstate-lock\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536336 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-ovs-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536421 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-nmstate-lock\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536480 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-ovs-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.536529 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b51a4da1-a921-414f-9985-044264a489ad-dbus-socket\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.550613 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.555663 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-58fcf6886f-9x5tk"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.556552 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.567872 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58fcf6886f-9x5tk"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.572235 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxt2g\" (UniqueName: \"kubernetes.io/projected/b51a4da1-a921-414f-9985-044264a489ad-kube-api-access-kxt2g\") pod \"nmstate-handler-wxpbs\" (UID: \"b51a4da1-a921-414f-9985-044264a489ad\") " pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.576107 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.637245 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hspdq\" (UniqueName: \"kubernetes.io/projected/d5575bcb-0d29-4538-8213-99a007548880-kube-api-access-hspdq\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.637294 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5575bcb-0d29-4538-8213-99a007548880-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.637388 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5575bcb-0d29-4538-8213-99a007548880-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.638431 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d5575bcb-0d29-4538-8213-99a007548880-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.641418 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d5575bcb-0d29-4538-8213-99a007548880-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.653934 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hspdq\" (UniqueName: \"kubernetes.io/projected/d5575bcb-0d29-4538-8213-99a007548880-kube-api-access-hspdq\") pod \"nmstate-console-plugin-7fbb5f6569-tjd2c\" (UID: \"d5575bcb-0d29-4538-8213-99a007548880\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.681144 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738460 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-oauth-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738512 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738532 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-service-ca\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738572 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfkrc\" (UniqueName: \"kubernetes.io/projected/aced32e0-77e6-4e82-a7cc-4410152e99c0-kube-api-access-wfkrc\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738587 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738626 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-trusted-ca-bundle\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.738648 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-oauth-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839618 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfkrc\" (UniqueName: \"kubernetes.io/projected/aced32e0-77e6-4e82-a7cc-4410152e99c0-kube-api-access-wfkrc\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839664 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839723 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-trusted-ca-bundle\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839747 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-oauth-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839785 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-oauth-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839807 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.839825 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-service-ca\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.840701 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-service-ca\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.840706 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-oauth-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.842866 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-trusted-ca-bundle\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.842947 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.847394 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-serving-cert\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.847410 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aced32e0-77e6-4e82-a7cc-4410152e99c0-console-oauth-config\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.857627 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfkrc\" (UniqueName: \"kubernetes.io/projected/aced32e0-77e6-4e82-a7cc-4410152e99c0-kube-api-access-wfkrc\") pod \"console-58fcf6886f-9x5tk\" (UID: \"aced32e0-77e6-4e82-a7cc-4410152e99c0\") " pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.889965 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c"] Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.906739 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wxpbs" event={"ID":"b51a4da1-a921-414f-9985-044264a489ad","Type":"ContainerStarted","Data":"c384a1e1e54b16b9bd674b54fb93518963f61ca44b88d4edb3ba9ed16a02f5d3"} Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.908286 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" event={"ID":"d5575bcb-0d29-4538-8213-99a007548880","Type":"ContainerStarted","Data":"da046e9c61f1651127f07a944f575b06016c4c1b8245ed38e4f45623bc48a7d7"} Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.941681 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.944100 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:44 crc kubenswrapper[4847]: I1210 14:36:44.946877 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ef0ffd00-af07-4843-9b34-2cf613252d7c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4kbcp\" (UID: \"ef0ffd00-af07-4843-9b34-2cf613252d7c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.021382 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v"] Dec 10 14:36:45 crc kubenswrapper[4847]: W1210 14:36:45.025110 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7bb30c3_5f7b_465a_8852_388f8204277b.slice/crio-2b78d7a13982bc6df2405d166d79b7fbbf580bf6c1bacf0657a460998c21c151 WatchSource:0}: Error finding container 2b78d7a13982bc6df2405d166d79b7fbbf580bf6c1bacf0657a460998c21c151: Status 404 returned error can't find the container with id 2b78d7a13982bc6df2405d166d79b7fbbf580bf6c1bacf0657a460998c21c151 Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.148385 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58fcf6886f-9x5tk"] Dec 10 14:36:45 crc kubenswrapper[4847]: W1210 14:36:45.153310 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaced32e0_77e6_4e82_a7cc_4410152e99c0.slice/crio-69e16ac4aa5e6ce84bb5b744946e29b0b6f98412ce0c5a030f738868b3a3d219 WatchSource:0}: Error finding container 69e16ac4aa5e6ce84bb5b744946e29b0b6f98412ce0c5a030f738868b3a3d219: Status 404 returned error can't find the container with id 69e16ac4aa5e6ce84bb5b744946e29b0b6f98412ce0c5a030f738868b3a3d219 Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.161825 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.356651 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp"] Dec 10 14:36:45 crc kubenswrapper[4847]: W1210 14:36:45.359210 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef0ffd00_af07_4843_9b34_2cf613252d7c.slice/crio-c105c7e3403317e5a729ae6232164a172450a673cfd0157628fd91ce57c758cd WatchSource:0}: Error finding container c105c7e3403317e5a729ae6232164a172450a673cfd0157628fd91ce57c758cd: Status 404 returned error can't find the container with id c105c7e3403317e5a729ae6232164a172450a673cfd0157628fd91ce57c758cd Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.916038 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58fcf6886f-9x5tk" event={"ID":"aced32e0-77e6-4e82-a7cc-4410152e99c0","Type":"ContainerStarted","Data":"0e15443b1441ddd3bff6eb508ca07d045ba625c109878f75a1f6645d15660738"} Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.916123 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58fcf6886f-9x5tk" event={"ID":"aced32e0-77e6-4e82-a7cc-4410152e99c0","Type":"ContainerStarted","Data":"69e16ac4aa5e6ce84bb5b744946e29b0b6f98412ce0c5a030f738868b3a3d219"} Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.918400 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" event={"ID":"e7bb30c3-5f7b-465a-8852-388f8204277b","Type":"ContainerStarted","Data":"2b78d7a13982bc6df2405d166d79b7fbbf580bf6c1bacf0657a460998c21c151"} Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.921505 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" event={"ID":"ef0ffd00-af07-4843-9b34-2cf613252d7c","Type":"ContainerStarted","Data":"c105c7e3403317e5a729ae6232164a172450a673cfd0157628fd91ce57c758cd"} Dec 10 14:36:45 crc kubenswrapper[4847]: I1210 14:36:45.936268 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-58fcf6886f-9x5tk" podStartSLOduration=1.936244525 podStartE2EDuration="1.936244525s" podCreationTimestamp="2025-12-10 14:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:36:45.935422522 +0000 UTC m=+755.504640172" watchObservedRunningTime="2025-12-10 14:36:45.936244525 +0000 UTC m=+755.505462155" Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.935741 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wxpbs" event={"ID":"b51a4da1-a921-414f-9985-044264a489ad","Type":"ContainerStarted","Data":"6968adb1f3cffe7a12680aa71c26093986bb2243178a7a20263dc51fd9f5d335"} Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.936680 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.938104 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" event={"ID":"d5575bcb-0d29-4538-8213-99a007548880","Type":"ContainerStarted","Data":"824df280ea6fa228a2344b83d29de18a079d2cd6ddf07f0d7afcb5a1ae5e32dc"} Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.939821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" event={"ID":"e7bb30c3-5f7b-465a-8852-388f8204277b","Type":"ContainerStarted","Data":"0aa03808e2bc2c92214ff9c5b6323d10b3e8131ec3fd708075f39907f75c5287"} Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.941477 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" event={"ID":"ef0ffd00-af07-4843-9b34-2cf613252d7c","Type":"ContainerStarted","Data":"835678378ef25e48282b1ef8bf3694aeda37238548b330ea5e60ef966db2570b"} Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.941627 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.958189 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wxpbs" podStartSLOduration=1.2718201740000001 podStartE2EDuration="3.958162171s" podCreationTimestamp="2025-12-10 14:36:44 +0000 UTC" firstStartedPulling="2025-12-10 14:36:44.612205828 +0000 UTC m=+754.181423458" lastFinishedPulling="2025-12-10 14:36:47.298547825 +0000 UTC m=+756.867765455" observedRunningTime="2025-12-10 14:36:47.953191513 +0000 UTC m=+757.522409163" watchObservedRunningTime="2025-12-10 14:36:47.958162171 +0000 UTC m=+757.527379801" Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.971314 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" podStartSLOduration=2.025305321 podStartE2EDuration="3.971283435s" podCreationTimestamp="2025-12-10 14:36:44 +0000 UTC" firstStartedPulling="2025-12-10 14:36:45.362108846 +0000 UTC m=+754.931326466" lastFinishedPulling="2025-12-10 14:36:47.30808695 +0000 UTC m=+756.877304580" observedRunningTime="2025-12-10 14:36:47.96820207 +0000 UTC m=+757.537419710" watchObservedRunningTime="2025-12-10 14:36:47.971283435 +0000 UTC m=+757.540501085" Dec 10 14:36:47 crc kubenswrapper[4847]: I1210 14:36:47.986937 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-tjd2c" podStartSLOduration=1.680590825 podStartE2EDuration="3.986919161s" podCreationTimestamp="2025-12-10 14:36:44 +0000 UTC" firstStartedPulling="2025-12-10 14:36:44.893513453 +0000 UTC m=+754.462731083" lastFinishedPulling="2025-12-10 14:36:47.199841789 +0000 UTC m=+756.769059419" observedRunningTime="2025-12-10 14:36:47.984269077 +0000 UTC m=+757.553486707" watchObservedRunningTime="2025-12-10 14:36:47.986919161 +0000 UTC m=+757.556136791" Dec 10 14:36:50 crc kubenswrapper[4847]: I1210 14:36:50.961728 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" event={"ID":"e7bb30c3-5f7b-465a-8852-388f8204277b","Type":"ContainerStarted","Data":"2e626f111ad4d3ec507f4e95b59ff4fdc94696277fd9db2a73dbf9a016044cfb"} Dec 10 14:36:50 crc kubenswrapper[4847]: I1210 14:36:50.978975 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-5rj2v" podStartSLOduration=1.996791847 podStartE2EDuration="6.978951028s" podCreationTimestamp="2025-12-10 14:36:44 +0000 UTC" firstStartedPulling="2025-12-10 14:36:45.028678502 +0000 UTC m=+754.597896132" lastFinishedPulling="2025-12-10 14:36:50.010837683 +0000 UTC m=+759.580055313" observedRunningTime="2025-12-10 14:36:50.976692985 +0000 UTC m=+760.545910625" watchObservedRunningTime="2025-12-10 14:36:50.978951028 +0000 UTC m=+760.548168658" Dec 10 14:36:54 crc kubenswrapper[4847]: I1210 14:36:54.600355 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wxpbs" Dec 10 14:36:54 crc kubenswrapper[4847]: I1210 14:36:54.945195 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:54 crc kubenswrapper[4847]: I1210 14:36:54.945295 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:54 crc kubenswrapper[4847]: I1210 14:36:54.949253 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:55 crc kubenswrapper[4847]: I1210 14:36:55.004689 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-58fcf6886f-9x5tk" Dec 10 14:36:55 crc kubenswrapper[4847]: I1210 14:36:55.058197 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:37:01 crc kubenswrapper[4847]: I1210 14:37:01.011373 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:37:01 crc kubenswrapper[4847]: I1210 14:37:01.012092 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:37:01 crc kubenswrapper[4847]: I1210 14:37:01.012194 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:37:01 crc kubenswrapper[4847]: I1210 14:37:01.012928 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:37:01 crc kubenswrapper[4847]: I1210 14:37:01.013007 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5" gracePeriod=600 Dec 10 14:37:02 crc kubenswrapper[4847]: I1210 14:37:02.042179 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5" exitCode=0 Dec 10 14:37:02 crc kubenswrapper[4847]: I1210 14:37:02.042518 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5"} Dec 10 14:37:02 crc kubenswrapper[4847]: I1210 14:37:02.042562 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006"} Dec 10 14:37:02 crc kubenswrapper[4847]: I1210 14:37:02.042578 4847 scope.go:117] "RemoveContainer" containerID="d42231ca2da354924be3c109cbb5e16d0881bab2788c509fbeb2b1870dc90b2d" Dec 10 14:37:05 crc kubenswrapper[4847]: I1210 14:37:05.168430 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4kbcp" Dec 10 14:37:18 crc kubenswrapper[4847]: I1210 14:37:18.990596 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm"] Dec 10 14:37:18 crc kubenswrapper[4847]: I1210 14:37:18.992564 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:18 crc kubenswrapper[4847]: I1210 14:37:18.994190 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.000265 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm"] Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.146001 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.146043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.146297 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhh5w\" (UniqueName: \"kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.248403 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhh5w\" (UniqueName: \"kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.248486 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.248516 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.249106 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.249454 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.274320 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhh5w\" (UniqueName: \"kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.310804 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:19 crc kubenswrapper[4847]: I1210 14:37:19.723502 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm"] Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.149262 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" event={"ID":"113a6c40-884a-41c9-9019-6ad88c99e989","Type":"ContainerStarted","Data":"2f463374cc30e0b1dbabd9d89e56cec4c725005b2dce195c0381b214cbd88790"} Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.169690 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-m2sr6" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" containerID="cri-o://a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70" gracePeriod=15 Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.595959 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m2sr6_24d0c394-4cf4-452f-a117-215657acb8f2/console/0.log" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.596021 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.719833 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:20 crc kubenswrapper[4847]: E1210 14:37:20.729763 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.729813 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.733236 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" containerName="console" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.736417 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.739458 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.771941 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.772765 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.772829 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.772914 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.773439 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca" (OuterVolumeSpecName: "service-ca") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.776830 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777046 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g8bg\" (UniqueName: \"kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777084 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777107 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777171 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert\") pod \"24d0c394-4cf4-452f-a117-215657acb8f2\" (UID: \"24d0c394-4cf4-452f-a117-215657acb8f2\") " Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777320 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trvlc\" (UniqueName: \"kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777374 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777424 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777464 4847 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777475 4847 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.777484 4847 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.778793 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config" (OuterVolumeSpecName: "console-config") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.785609 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg" (OuterVolumeSpecName: "kube-api-access-4g8bg") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "kube-api-access-4g8bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.787891 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.795903 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "24d0c394-4cf4-452f-a117-215657acb8f2" (UID: "24d0c394-4cf4-452f-a117-215657acb8f2"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878778 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878862 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878912 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trvlc\" (UniqueName: \"kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878953 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g8bg\" (UniqueName: \"kubernetes.io/projected/24d0c394-4cf4-452f-a117-215657acb8f2-kube-api-access-4g8bg\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878964 4847 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/24d0c394-4cf4-452f-a117-215657acb8f2-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878973 4847 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.878982 4847 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/24d0c394-4cf4-452f-a117-215657acb8f2-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.879421 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.879507 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:20 crc kubenswrapper[4847]: I1210 14:37:20.895576 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trvlc\" (UniqueName: \"kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc\") pod \"redhat-operators-qwm9b\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.074835 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.155981 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m2sr6_24d0c394-4cf4-452f-a117-215657acb8f2/console/0.log" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.156046 4847 generic.go:334] "Generic (PLEG): container finished" podID="24d0c394-4cf4-452f-a117-215657acb8f2" containerID="a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70" exitCode=2 Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.156108 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2sr6" event={"ID":"24d0c394-4cf4-452f-a117-215657acb8f2","Type":"ContainerDied","Data":"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70"} Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.156138 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2sr6" event={"ID":"24d0c394-4cf4-452f-a117-215657acb8f2","Type":"ContainerDied","Data":"207d1362d360284dbb7f224a9bfd265caf3c850c58b071ef7b4e5a84d072331b"} Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.156158 4847 scope.go:117] "RemoveContainer" containerID="a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.156290 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2sr6" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.175506 4847 generic.go:334] "Generic (PLEG): container finished" podID="113a6c40-884a-41c9-9019-6ad88c99e989" containerID="84da77e942ddea46837257975d1592e9d8973517161ac9e647ea865d475d4e50" exitCode=0 Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.175543 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" event={"ID":"113a6c40-884a-41c9-9019-6ad88c99e989","Type":"ContainerDied","Data":"84da77e942ddea46837257975d1592e9d8973517161ac9e647ea865d475d4e50"} Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.195195 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.206329 4847 scope.go:117] "RemoveContainer" containerID="a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70" Dec 10 14:37:21 crc kubenswrapper[4847]: E1210 14:37:21.206616 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70\": container with ID starting with a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70 not found: ID does not exist" containerID="a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.206661 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70"} err="failed to get container status \"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70\": rpc error: code = NotFound desc = could not find container \"a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70\": container with ID starting with a52a21aad3bbc996b00b623997fd0e17b4a92f31452ff59fbdc2c907bc74be70 not found: ID does not exist" Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.210675 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-m2sr6"] Dec 10 14:37:21 crc kubenswrapper[4847]: I1210 14:37:21.327266 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:21 crc kubenswrapper[4847]: W1210 14:37:21.340491 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6b49ff2_919c_422d_9780_09448e769d78.slice/crio-b975ab9493e57b8517582cdb577cb93b9458c6fdda4f4917548485281078fc7c WatchSource:0}: Error finding container b975ab9493e57b8517582cdb577cb93b9458c6fdda4f4917548485281078fc7c: Status 404 returned error can't find the container with id b975ab9493e57b8517582cdb577cb93b9458c6fdda4f4917548485281078fc7c Dec 10 14:37:22 crc kubenswrapper[4847]: I1210 14:37:22.183644 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6b49ff2-919c-422d-9780-09448e769d78" containerID="549f8f34ccc3a5cbf51847d189dbc72106ed2dfe425a8cc231d423b8e5cb5e64" exitCode=0 Dec 10 14:37:22 crc kubenswrapper[4847]: I1210 14:37:22.183742 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerDied","Data":"549f8f34ccc3a5cbf51847d189dbc72106ed2dfe425a8cc231d423b8e5cb5e64"} Dec 10 14:37:22 crc kubenswrapper[4847]: I1210 14:37:22.184097 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerStarted","Data":"b975ab9493e57b8517582cdb577cb93b9458c6fdda4f4917548485281078fc7c"} Dec 10 14:37:22 crc kubenswrapper[4847]: I1210 14:37:22.769663 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d0c394-4cf4-452f-a117-215657acb8f2" path="/var/lib/kubelet/pods/24d0c394-4cf4-452f-a117-215657acb8f2/volumes" Dec 10 14:37:23 crc kubenswrapper[4847]: I1210 14:37:23.191279 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerStarted","Data":"7d1cafc2d85c2bec7342789111ac5054cd4c0854e68ebd7e522601795717c352"} Dec 10 14:37:23 crc kubenswrapper[4847]: I1210 14:37:23.193653 4847 generic.go:334] "Generic (PLEG): container finished" podID="113a6c40-884a-41c9-9019-6ad88c99e989" containerID="4d2b5775d93cc68e987ca1e60bd029f6dc1611ec59937bf17d7eb5b66a3c8f45" exitCode=0 Dec 10 14:37:23 crc kubenswrapper[4847]: I1210 14:37:23.193722 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" event={"ID":"113a6c40-884a-41c9-9019-6ad88c99e989","Type":"ContainerDied","Data":"4d2b5775d93cc68e987ca1e60bd029f6dc1611ec59937bf17d7eb5b66a3c8f45"} Dec 10 14:37:24 crc kubenswrapper[4847]: I1210 14:37:24.201646 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6b49ff2-919c-422d-9780-09448e769d78" containerID="7d1cafc2d85c2bec7342789111ac5054cd4c0854e68ebd7e522601795717c352" exitCode=0 Dec 10 14:37:24 crc kubenswrapper[4847]: I1210 14:37:24.201706 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerDied","Data":"7d1cafc2d85c2bec7342789111ac5054cd4c0854e68ebd7e522601795717c352"} Dec 10 14:37:24 crc kubenswrapper[4847]: I1210 14:37:24.204389 4847 generic.go:334] "Generic (PLEG): container finished" podID="113a6c40-884a-41c9-9019-6ad88c99e989" containerID="610632f5961f99009733757b32d49905660685dfc408515120eabcda093cb04d" exitCode=0 Dec 10 14:37:24 crc kubenswrapper[4847]: I1210 14:37:24.204448 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" event={"ID":"113a6c40-884a-41c9-9019-6ad88c99e989","Type":"ContainerDied","Data":"610632f5961f99009733757b32d49905660685dfc408515120eabcda093cb04d"} Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.212910 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerStarted","Data":"3965ee2a56379210f8a31e3f47298e85c887b71bd12b1e38f7af281dee3d7700"} Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.241776 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qwm9b" podStartSLOduration=2.764381506 podStartE2EDuration="5.241755069s" podCreationTimestamp="2025-12-10 14:37:20 +0000 UTC" firstStartedPulling="2025-12-10 14:37:22.185115084 +0000 UTC m=+791.754332714" lastFinishedPulling="2025-12-10 14:37:24.662488637 +0000 UTC m=+794.231706277" observedRunningTime="2025-12-10 14:37:25.235891866 +0000 UTC m=+794.805109496" watchObservedRunningTime="2025-12-10 14:37:25.241755069 +0000 UTC m=+794.810972699" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.458039 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.551758 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle\") pod \"113a6c40-884a-41c9-9019-6ad88c99e989\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.551846 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhh5w\" (UniqueName: \"kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w\") pod \"113a6c40-884a-41c9-9019-6ad88c99e989\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.551878 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util\") pod \"113a6c40-884a-41c9-9019-6ad88c99e989\" (UID: \"113a6c40-884a-41c9-9019-6ad88c99e989\") " Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.552902 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle" (OuterVolumeSpecName: "bundle") pod "113a6c40-884a-41c9-9019-6ad88c99e989" (UID: "113a6c40-884a-41c9-9019-6ad88c99e989"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.561301 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w" (OuterVolumeSpecName: "kube-api-access-jhh5w") pod "113a6c40-884a-41c9-9019-6ad88c99e989" (UID: "113a6c40-884a-41c9-9019-6ad88c99e989"). InnerVolumeSpecName "kube-api-access-jhh5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.653298 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhh5w\" (UniqueName: \"kubernetes.io/projected/113a6c40-884a-41c9-9019-6ad88c99e989-kube-api-access-jhh5w\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.653356 4847 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.685098 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util" (OuterVolumeSpecName: "util") pod "113a6c40-884a-41c9-9019-6ad88c99e989" (UID: "113a6c40-884a-41c9-9019-6ad88c99e989"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:37:25 crc kubenswrapper[4847]: I1210 14:37:25.754377 4847 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/113a6c40-884a-41c9-9019-6ad88c99e989-util\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:26 crc kubenswrapper[4847]: I1210 14:37:26.222273 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" event={"ID":"113a6c40-884a-41c9-9019-6ad88c99e989","Type":"ContainerDied","Data":"2f463374cc30e0b1dbabd9d89e56cec4c725005b2dce195c0381b214cbd88790"} Dec 10 14:37:26 crc kubenswrapper[4847]: I1210 14:37:26.222358 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm" Dec 10 14:37:26 crc kubenswrapper[4847]: I1210 14:37:26.222358 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f463374cc30e0b1dbabd9d89e56cec4c725005b2dce195c0381b214cbd88790" Dec 10 14:37:31 crc kubenswrapper[4847]: I1210 14:37:31.075197 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:31 crc kubenswrapper[4847]: I1210 14:37:31.075687 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:31 crc kubenswrapper[4847]: I1210 14:37:31.117083 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:31 crc kubenswrapper[4847]: I1210 14:37:31.333867 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:33 crc kubenswrapper[4847]: I1210 14:37:33.102697 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:33 crc kubenswrapper[4847]: I1210 14:37:33.281546 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qwm9b" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="registry-server" containerID="cri-o://3965ee2a56379210f8a31e3f47298e85c887b71bd12b1e38f7af281dee3d7700" gracePeriod=2 Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.311002 4847 generic.go:334] "Generic (PLEG): container finished" podID="a6b49ff2-919c-422d-9780-09448e769d78" containerID="3965ee2a56379210f8a31e3f47298e85c887b71bd12b1e38f7af281dee3d7700" exitCode=0 Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.311150 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerDied","Data":"3965ee2a56379210f8a31e3f47298e85c887b71bd12b1e38f7af281dee3d7700"} Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.408510 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.443474 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd"] Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444000 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="extract-utilities" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444097 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="extract-utilities" Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444169 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="extract-content" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444231 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="extract-content" Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444296 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="extract" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444356 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="extract" Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444432 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="pull" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444492 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="pull" Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444566 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="registry-server" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444635 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="registry-server" Dec 10 14:37:37 crc kubenswrapper[4847]: E1210 14:37:37.444705 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="util" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444788 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="util" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.444975 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="113a6c40-884a-41c9-9019-6ad88c99e989" containerName="extract" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.445066 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6b49ff2-919c-422d-9780-09448e769d78" containerName="registry-server" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.445738 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.453476 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.454967 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-chjgp" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.461001 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.461023 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.461044 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.479736 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd"] Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604117 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trvlc\" (UniqueName: \"kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc\") pod \"a6b49ff2-919c-422d-9780-09448e769d78\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604265 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities\") pod \"a6b49ff2-919c-422d-9780-09448e769d78\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604337 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content\") pod \"a6b49ff2-919c-422d-9780-09448e769d78\" (UID: \"a6b49ff2-919c-422d-9780-09448e769d78\") " Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604533 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-apiservice-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604575 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-webhook-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.604621 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96cjz\" (UniqueName: \"kubernetes.io/projected/12bcc429-a2dd-4b92-a99b-86ea695e6147-kube-api-access-96cjz\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.605092 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities" (OuterVolumeSpecName: "utilities") pod "a6b49ff2-919c-422d-9780-09448e769d78" (UID: "a6b49ff2-919c-422d-9780-09448e769d78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.613405 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc" (OuterVolumeSpecName: "kube-api-access-trvlc") pod "a6b49ff2-919c-422d-9780-09448e769d78" (UID: "a6b49ff2-919c-422d-9780-09448e769d78"). InnerVolumeSpecName "kube-api-access-trvlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.705941 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-apiservice-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.706013 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-webhook-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.706076 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96cjz\" (UniqueName: \"kubernetes.io/projected/12bcc429-a2dd-4b92-a99b-86ea695e6147-kube-api-access-96cjz\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.706125 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trvlc\" (UniqueName: \"kubernetes.io/projected/a6b49ff2-919c-422d-9780-09448e769d78-kube-api-access-trvlc\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.706142 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.710069 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-apiservice-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.714850 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12bcc429-a2dd-4b92-a99b-86ea695e6147-webhook-cert\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.720101 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6b49ff2-919c-422d-9780-09448e769d78" (UID: "a6b49ff2-919c-422d-9780-09448e769d78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.725513 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96cjz\" (UniqueName: \"kubernetes.io/projected/12bcc429-a2dd-4b92-a99b-86ea695e6147-kube-api-access-96cjz\") pod \"metallb-operator-controller-manager-5b5f976d54-lrrhd\" (UID: \"12bcc429-a2dd-4b92-a99b-86ea695e6147\") " pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.760796 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.785657 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv"] Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.786385 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.789685 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jdrhr" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.789809 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.789816 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.807989 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6b49ff2-919c-422d-9780-09448e769d78-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.824849 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv"] Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.909707 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-apiservice-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.909773 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-webhook-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:37 crc kubenswrapper[4847]: I1210 14:37:37.909819 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4vf5\" (UniqueName: \"kubernetes.io/projected/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-kube-api-access-w4vf5\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.011608 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4vf5\" (UniqueName: \"kubernetes.io/projected/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-kube-api-access-w4vf5\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.012089 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-apiservice-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.012116 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-webhook-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.017507 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-apiservice-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.023528 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd"] Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.024330 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-webhook-cert\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.033321 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4vf5\" (UniqueName: \"kubernetes.io/projected/528aeb2a-2b6f-4e7b-8a22-b40ecca39c45-kube-api-access-w4vf5\") pod \"metallb-operator-webhook-server-7bc66b7c4c-gxscv\" (UID: \"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45\") " pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.118010 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.325594 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwm9b" event={"ID":"a6b49ff2-919c-422d-9780-09448e769d78","Type":"ContainerDied","Data":"b975ab9493e57b8517582cdb577cb93b9458c6fdda4f4917548485281078fc7c"} Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.325665 4847 scope.go:117] "RemoveContainer" containerID="3965ee2a56379210f8a31e3f47298e85c887b71bd12b1e38f7af281dee3d7700" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.325693 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwm9b" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.327508 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" event={"ID":"12bcc429-a2dd-4b92-a99b-86ea695e6147","Type":"ContainerStarted","Data":"05325c5716460b586fb322dfd2e9eb6db5111ab711ce1b7de96c4a0521b8de3b"} Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.335471 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv"] Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.359478 4847 scope.go:117] "RemoveContainer" containerID="7d1cafc2d85c2bec7342789111ac5054cd4c0854e68ebd7e522601795717c352" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.368023 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.372427 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qwm9b"] Dec 10 14:37:38 crc kubenswrapper[4847]: W1210 14:37:38.373562 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod528aeb2a_2b6f_4e7b_8a22_b40ecca39c45.slice/crio-f5e55c3e455a0356489c70d519593580b145c7917911cfe1ef6dbe0ad16137fe WatchSource:0}: Error finding container f5e55c3e455a0356489c70d519593580b145c7917911cfe1ef6dbe0ad16137fe: Status 404 returned error can't find the container with id f5e55c3e455a0356489c70d519593580b145c7917911cfe1ef6dbe0ad16137fe Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.417837 4847 scope.go:117] "RemoveContainer" containerID="549f8f34ccc3a5cbf51847d189dbc72106ed2dfe425a8cc231d423b8e5cb5e64" Dec 10 14:37:38 crc kubenswrapper[4847]: I1210 14:37:38.767349 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6b49ff2-919c-422d-9780-09448e769d78" path="/var/lib/kubelet/pods/a6b49ff2-919c-422d-9780-09448e769d78/volumes" Dec 10 14:37:39 crc kubenswrapper[4847]: I1210 14:37:39.333333 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" event={"ID":"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45","Type":"ContainerStarted","Data":"f5e55c3e455a0356489c70d519593580b145c7917911cfe1ef6dbe0ad16137fe"} Dec 10 14:37:41 crc kubenswrapper[4847]: I1210 14:37:41.358599 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" event={"ID":"12bcc429-a2dd-4b92-a99b-86ea695e6147","Type":"ContainerStarted","Data":"8d27227a6d0026cbde3951f9b6b641b6c19ac59ef1bf2ea6df490a6badf226f4"} Dec 10 14:37:41 crc kubenswrapper[4847]: I1210 14:37:41.360057 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:37:41 crc kubenswrapper[4847]: I1210 14:37:41.395397 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" podStartSLOduration=1.366308515 podStartE2EDuration="4.395375363s" podCreationTimestamp="2025-12-10 14:37:37 +0000 UTC" firstStartedPulling="2025-12-10 14:37:38.039391392 +0000 UTC m=+807.608609022" lastFinishedPulling="2025-12-10 14:37:41.06845824 +0000 UTC m=+810.637675870" observedRunningTime="2025-12-10 14:37:41.392590716 +0000 UTC m=+810.961808336" watchObservedRunningTime="2025-12-10 14:37:41.395375363 +0000 UTC m=+810.964593033" Dec 10 14:37:43 crc kubenswrapper[4847]: I1210 14:37:43.372417 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" event={"ID":"528aeb2a-2b6f-4e7b-8a22-b40ecca39c45","Type":"ContainerStarted","Data":"5b391c25aee1301aa32aedfb3994a85f32ec4d107c07f9cb257efd849c57368a"} Dec 10 14:37:43 crc kubenswrapper[4847]: I1210 14:37:43.373358 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:37:43 crc kubenswrapper[4847]: I1210 14:37:43.392567 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" podStartSLOduration=1.964508953 podStartE2EDuration="6.392542811s" podCreationTimestamp="2025-12-10 14:37:37 +0000 UTC" firstStartedPulling="2025-12-10 14:37:38.378203406 +0000 UTC m=+807.947421036" lastFinishedPulling="2025-12-10 14:37:42.806237264 +0000 UTC m=+812.375454894" observedRunningTime="2025-12-10 14:37:43.388433797 +0000 UTC m=+812.957651437" watchObservedRunningTime="2025-12-10 14:37:43.392542811 +0000 UTC m=+812.961760441" Dec 10 14:37:58 crc kubenswrapper[4847]: I1210 14:37:58.123983 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7bc66b7c4c-gxscv" Dec 10 14:38:17 crc kubenswrapper[4847]: I1210 14:38:17.764507 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b5f976d54-lrrhd" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.478557 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-96l7c"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.480924 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.483207 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.483489 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.483504 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-f9rtn" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.490979 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.492163 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.494857 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.497865 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.563933 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-reloader\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.564635 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.564789 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-sockets\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.564912 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-conf\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.565012 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics-certs\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.565103 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn6fn\" (UniqueName: \"kubernetes.io/projected/b837e40a-6332-4206-94f5-2555dc1cd40c-kube-api-access-xn6fn\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.565204 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-startup\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.565303 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec15bd80-1069-4aed-9f51-43140a7f6c5e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.565420 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56l8w\" (UniqueName: \"kubernetes.io/projected/ec15bd80-1069-4aed-9f51-43140a7f6c5e-kube-api-access-56l8w\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.564655 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7bf6v"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.566730 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.571384 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5v8z2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.571805 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.572012 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.572243 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.578796 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-6mjl9"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.580009 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.582070 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.596640 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6mjl9"] Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666703 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-metrics-certs\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666756 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666813 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666848 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-sockets\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666886 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-conf\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666915 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics-certs\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666937 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-metrics-certs\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666958 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn6fn\" (UniqueName: \"kubernetes.io/projected/b837e40a-6332-4206-94f5-2555dc1cd40c-kube-api-access-xn6fn\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.666980 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-startup\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667001 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwg4w\" (UniqueName: \"kubernetes.io/projected/74b7d233-5faf-4997-9986-f6fb167df94b-kube-api-access-fwg4w\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667027 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec15bd80-1069-4aed-9f51-43140a7f6c5e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667120 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-cert\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667163 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/74b7d233-5faf-4997-9986-f6fb167df94b-metallb-excludel2\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667189 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56l8w\" (UniqueName: \"kubernetes.io/projected/ec15bd80-1069-4aed-9f51-43140a7f6c5e-kube-api-access-56l8w\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667237 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd5cc\" (UniqueName: \"kubernetes.io/projected/d00cc0ff-5b61-4786-abd8-44da90bdb484-kube-api-access-zd5cc\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667264 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-reloader\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667382 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667423 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-sockets\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.667827 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-reloader\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.668069 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-conf\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.668734 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b837e40a-6332-4206-94f5-2555dc1cd40c-frr-startup\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.676879 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b837e40a-6332-4206-94f5-2555dc1cd40c-metrics-certs\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.682737 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ec15bd80-1069-4aed-9f51-43140a7f6c5e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.687858 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn6fn\" (UniqueName: \"kubernetes.io/projected/b837e40a-6332-4206-94f5-2555dc1cd40c-kube-api-access-xn6fn\") pod \"frr-k8s-96l7c\" (UID: \"b837e40a-6332-4206-94f5-2555dc1cd40c\") " pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.690057 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56l8w\" (UniqueName: \"kubernetes.io/projected/ec15bd80-1069-4aed-9f51-43140a7f6c5e-kube-api-access-56l8w\") pod \"frr-k8s-webhook-server-7fcb986d4-ng6j2\" (UID: \"ec15bd80-1069-4aed-9f51-43140a7f6c5e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768485 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-metrics-certs\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768545 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwg4w\" (UniqueName: \"kubernetes.io/projected/74b7d233-5faf-4997-9986-f6fb167df94b-kube-api-access-fwg4w\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768584 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-cert\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768612 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/74b7d233-5faf-4997-9986-f6fb167df94b-metallb-excludel2\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768645 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd5cc\" (UniqueName: \"kubernetes.io/projected/d00cc0ff-5b61-4786-abd8-44da90bdb484-kube-api-access-zd5cc\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768679 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-metrics-certs\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.768750 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: E1210 14:38:18.768895 4847 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 14:38:18 crc kubenswrapper[4847]: E1210 14:38:18.768955 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist podName:74b7d233-5faf-4997-9986-f6fb167df94b nodeName:}" failed. No retries permitted until 2025-12-10 14:38:19.268934134 +0000 UTC m=+848.838151764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist") pod "speaker-7bf6v" (UID: "74b7d233-5faf-4997-9986-f6fb167df94b") : secret "metallb-memberlist" not found Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.770218 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/74b7d233-5faf-4997-9986-f6fb167df94b-metallb-excludel2\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.771962 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-metrics-certs\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.772370 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-metrics-certs\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.773891 4847 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.786938 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwg4w\" (UniqueName: \"kubernetes.io/projected/74b7d233-5faf-4997-9986-f6fb167df94b-kube-api-access-fwg4w\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.788682 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd5cc\" (UniqueName: \"kubernetes.io/projected/d00cc0ff-5b61-4786-abd8-44da90bdb484-kube-api-access-zd5cc\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.793302 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d00cc0ff-5b61-4786-abd8-44da90bdb484-cert\") pod \"controller-f8648f98b-6mjl9\" (UID: \"d00cc0ff-5b61-4786-abd8-44da90bdb484\") " pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.799305 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.806403 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:18 crc kubenswrapper[4847]: I1210 14:38:18.893988 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.211658 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2"] Dec 10 14:38:19 crc kubenswrapper[4847]: W1210 14:38:19.222981 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec15bd80_1069_4aed_9f51_43140a7f6c5e.slice/crio-64b6e1c260767025e488a7e7fdead0baff83493b2a81c062f834709e34257fa6 WatchSource:0}: Error finding container 64b6e1c260767025e488a7e7fdead0baff83493b2a81c062f834709e34257fa6: Status 404 returned error can't find the container with id 64b6e1c260767025e488a7e7fdead0baff83493b2a81c062f834709e34257fa6 Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.275001 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:19 crc kubenswrapper[4847]: E1210 14:38:19.275226 4847 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 14:38:19 crc kubenswrapper[4847]: E1210 14:38:19.275330 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist podName:74b7d233-5faf-4997-9986-f6fb167df94b nodeName:}" failed. No retries permitted until 2025-12-10 14:38:20.275308738 +0000 UTC m=+849.844526378 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist") pod "speaker-7bf6v" (UID: "74b7d233-5faf-4997-9986-f6fb167df94b") : secret "metallb-memberlist" not found Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.330301 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6mjl9"] Dec 10 14:38:19 crc kubenswrapper[4847]: W1210 14:38:19.336474 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd00cc0ff_5b61_4786_abd8_44da90bdb484.slice/crio-93eb3ac7d88fd3fbfdbc26aaa073ee1fa927839ba200df623983e344004a303c WatchSource:0}: Error finding container 93eb3ac7d88fd3fbfdbc26aaa073ee1fa927839ba200df623983e344004a303c: Status 404 returned error can't find the container with id 93eb3ac7d88fd3fbfdbc26aaa073ee1fa927839ba200df623983e344004a303c Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.689502 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" event={"ID":"ec15bd80-1069-4aed-9f51-43140a7f6c5e","Type":"ContainerStarted","Data":"64b6e1c260767025e488a7e7fdead0baff83493b2a81c062f834709e34257fa6"} Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.694903 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6mjl9" event={"ID":"d00cc0ff-5b61-4786-abd8-44da90bdb484","Type":"ContainerStarted","Data":"bd032b76e5d3a04e13b1f8d282ee5097ba57ff190493c848bf2c928b1212d776"} Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.694950 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6mjl9" event={"ID":"d00cc0ff-5b61-4786-abd8-44da90bdb484","Type":"ContainerStarted","Data":"f4e2fa08ee6847748724852729824f399aba7ae9f027236c6a0c4ab29820c5f7"} Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.694963 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6mjl9" event={"ID":"d00cc0ff-5b61-4786-abd8-44da90bdb484","Type":"ContainerStarted","Data":"93eb3ac7d88fd3fbfdbc26aaa073ee1fa927839ba200df623983e344004a303c"} Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.695012 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:19 crc kubenswrapper[4847]: I1210 14:38:19.700456 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"b07f4dfe0e746310710b89a72c92fc053f563b83e7e5a440127028b89f655f50"} Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.288990 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.306374 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/74b7d233-5faf-4997-9986-f6fb167df94b-memberlist\") pod \"speaker-7bf6v\" (UID: \"74b7d233-5faf-4997-9986-f6fb167df94b\") " pod="metallb-system/speaker-7bf6v" Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.383448 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7bf6v" Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.710054 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7bf6v" event={"ID":"74b7d233-5faf-4997-9986-f6fb167df94b","Type":"ContainerStarted","Data":"a76ac27ff95bef2e55499fbaa8d547ecc8fbf9d2fc3c12c7d61eb84aaeefc94e"} Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.710319 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7bf6v" event={"ID":"74b7d233-5faf-4997-9986-f6fb167df94b","Type":"ContainerStarted","Data":"c5c6f64c2dbe1734957b6d87efc29bb852057b9113e03a4cd56ec3afa19f6d17"} Dec 10 14:38:20 crc kubenswrapper[4847]: I1210 14:38:20.782301 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-6mjl9" podStartSLOduration=2.782282623 podStartE2EDuration="2.782282623s" podCreationTimestamp="2025-12-10 14:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:38:19.716768737 +0000 UTC m=+849.285986377" watchObservedRunningTime="2025-12-10 14:38:20.782282623 +0000 UTC m=+850.351500253" Dec 10 14:38:21 crc kubenswrapper[4847]: I1210 14:38:21.723700 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7bf6v" event={"ID":"74b7d233-5faf-4997-9986-f6fb167df94b","Type":"ContainerStarted","Data":"a884310f6cfcc0c3dcd92acc678a8931270011b1298b48eed146348d6af19b1c"} Dec 10 14:38:21 crc kubenswrapper[4847]: I1210 14:38:21.724124 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7bf6v" Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.753100 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" event={"ID":"ec15bd80-1069-4aed-9f51-43140a7f6c5e","Type":"ContainerStarted","Data":"6be68e10b9d7c014b5da1968b24120f190813d90d796509398b34be9b29e00af"} Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.753669 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.756282 4847 generic.go:334] "Generic (PLEG): container finished" podID="b837e40a-6332-4206-94f5-2555dc1cd40c" containerID="9db7f919399d06355443ac3447a488e39e149555fdce4acf411163da3680dba4" exitCode=0 Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.756333 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerDied","Data":"9db7f919399d06355443ac3447a488e39e149555fdce4acf411163da3680dba4"} Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.769370 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7bf6v" podStartSLOduration=8.769351714 podStartE2EDuration="8.769351714s" podCreationTimestamp="2025-12-10 14:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:38:21.757027905 +0000 UTC m=+851.326245555" watchObservedRunningTime="2025-12-10 14:38:26.769351714 +0000 UTC m=+856.338569344" Dec 10 14:38:26 crc kubenswrapper[4847]: I1210 14:38:26.812147 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" podStartSLOduration=1.847833462 podStartE2EDuration="8.812128493s" podCreationTimestamp="2025-12-10 14:38:18 +0000 UTC" firstStartedPulling="2025-12-10 14:38:19.225068871 +0000 UTC m=+848.794286501" lastFinishedPulling="2025-12-10 14:38:26.189363902 +0000 UTC m=+855.758581532" observedRunningTime="2025-12-10 14:38:26.773663094 +0000 UTC m=+856.342880724" watchObservedRunningTime="2025-12-10 14:38:26.812128493 +0000 UTC m=+856.381346123" Dec 10 14:38:27 crc kubenswrapper[4847]: I1210 14:38:27.772786 4847 generic.go:334] "Generic (PLEG): container finished" podID="b837e40a-6332-4206-94f5-2555dc1cd40c" containerID="294c61605896f0d4ba8b6ea8f23283dd549dd5c24118d43d8f1d33a23b001339" exitCode=0 Dec 10 14:38:27 crc kubenswrapper[4847]: I1210 14:38:27.772838 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerDied","Data":"294c61605896f0d4ba8b6ea8f23283dd549dd5c24118d43d8f1d33a23b001339"} Dec 10 14:38:28 crc kubenswrapper[4847]: I1210 14:38:28.779986 4847 generic.go:334] "Generic (PLEG): container finished" podID="b837e40a-6332-4206-94f5-2555dc1cd40c" containerID="78678e6a7505bb717cce7dfc8ca879cba7292c5184e6fc54faf67b0c264f1c82" exitCode=0 Dec 10 14:38:28 crc kubenswrapper[4847]: I1210 14:38:28.780074 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerDied","Data":"78678e6a7505bb717cce7dfc8ca879cba7292c5184e6fc54faf67b0c264f1c82"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.789534 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"718b9c56b366a38fb6b8981ecb97e3e0f2584269bb1f9b7c48cd4125d9066c9a"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790414 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790437 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"e61466d1f6be03d716b9a5ef2a9b1056233aee208de9ec788e1e5b127e9e8124"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790450 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"499cc1f84beddbfb993055e7c115e1872169cab085a23e9a5bd8405e5d3b9d1d"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790465 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"21b32457d3d16bec1e0a5e7ddb94c6f02009b4c310bcf7973360ccc383b0adf3"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790476 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"54121cedae8b385422ab133ac9bb0c155726cd8dbdea842fcd5b41a2950c3a9c"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.790844 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-96l7c" event={"ID":"b837e40a-6332-4206-94f5-2555dc1cd40c","Type":"ContainerStarted","Data":"3c49fb3f6c9e74614288f100fb353f323a104eaf9f78396b69bc667aa5daed52"} Dec 10 14:38:29 crc kubenswrapper[4847]: I1210 14:38:29.813547 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-96l7c" podStartSLOduration=4.575487768 podStartE2EDuration="11.813527263s" podCreationTimestamp="2025-12-10 14:38:18 +0000 UTC" firstStartedPulling="2025-12-10 14:38:18.939621332 +0000 UTC m=+848.508838962" lastFinishedPulling="2025-12-10 14:38:26.177660827 +0000 UTC m=+855.746878457" observedRunningTime="2025-12-10 14:38:29.809452239 +0000 UTC m=+859.378669869" watchObservedRunningTime="2025-12-10 14:38:29.813527263 +0000 UTC m=+859.382744893" Dec 10 14:38:30 crc kubenswrapper[4847]: I1210 14:38:30.387601 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7bf6v" Dec 10 14:38:33 crc kubenswrapper[4847]: I1210 14:38:33.800014 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:33 crc kubenswrapper[4847]: I1210 14:38:33.844522 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.654383 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ptjmg"] Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.655914 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.658556 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.658837 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-qzxjf" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.663908 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.670886 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ptjmg"] Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.725841 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2spv\" (UniqueName: \"kubernetes.io/projected/692f8962-1682-4a0d-b693-24e90e81bb17-kube-api-access-z2spv\") pod \"openstack-operator-index-ptjmg\" (UID: \"692f8962-1682-4a0d-b693-24e90e81bb17\") " pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.826599 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2spv\" (UniqueName: \"kubernetes.io/projected/692f8962-1682-4a0d-b693-24e90e81bb17-kube-api-access-z2spv\") pod \"openstack-operator-index-ptjmg\" (UID: \"692f8962-1682-4a0d-b693-24e90e81bb17\") " pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.847086 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2spv\" (UniqueName: \"kubernetes.io/projected/692f8962-1682-4a0d-b693-24e90e81bb17-kube-api-access-z2spv\") pod \"openstack-operator-index-ptjmg\" (UID: \"692f8962-1682-4a0d-b693-24e90e81bb17\") " pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:36 crc kubenswrapper[4847]: I1210 14:38:36.974043 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:37 crc kubenswrapper[4847]: I1210 14:38:37.203244 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ptjmg"] Dec 10 14:38:37 crc kubenswrapper[4847]: I1210 14:38:37.836335 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ptjmg" event={"ID":"692f8962-1682-4a0d-b693-24e90e81bb17","Type":"ContainerStarted","Data":"739657197a8ac00a99116d147f35ebd76f11d495c1bc9d1937c1977a7b94861f"} Dec 10 14:38:38 crc kubenswrapper[4847]: I1210 14:38:38.802848 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-96l7c" Dec 10 14:38:38 crc kubenswrapper[4847]: I1210 14:38:38.840943 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-ng6j2" Dec 10 14:38:38 crc kubenswrapper[4847]: I1210 14:38:38.898419 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-6mjl9" Dec 10 14:38:40 crc kubenswrapper[4847]: I1210 14:38:40.858844 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ptjmg" event={"ID":"692f8962-1682-4a0d-b693-24e90e81bb17","Type":"ContainerStarted","Data":"48403f83cc6f58b4650b6ab0a1a8996ad1a6107582f72663e4ee7e3906a87f89"} Dec 10 14:38:40 crc kubenswrapper[4847]: I1210 14:38:40.873956 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ptjmg" podStartSLOduration=1.4138612130000001 podStartE2EDuration="4.873936479s" podCreationTimestamp="2025-12-10 14:38:36 +0000 UTC" firstStartedPulling="2025-12-10 14:38:37.210783764 +0000 UTC m=+866.780001394" lastFinishedPulling="2025-12-10 14:38:40.67085903 +0000 UTC m=+870.240076660" observedRunningTime="2025-12-10 14:38:40.872616492 +0000 UTC m=+870.441834142" watchObservedRunningTime="2025-12-10 14:38:40.873936479 +0000 UTC m=+870.443154109" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.259887 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.263450 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.280016 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.311589 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.311680 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vbl\" (UniqueName: \"kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.311726 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.412341 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vbl\" (UniqueName: \"kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.412388 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.412463 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.412992 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.413041 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.433789 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vbl\" (UniqueName: \"kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl\") pod \"redhat-marketplace-jbhg8\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.589335 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.817460 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:43 crc kubenswrapper[4847]: I1210 14:38:43.875396 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerStarted","Data":"438ff6e83ffa7b06912098160d04654e41df6f4c049b09ab281829398690a93d"} Dec 10 14:38:44 crc kubenswrapper[4847]: I1210 14:38:44.882935 4847 generic.go:334] "Generic (PLEG): container finished" podID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerID="9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd" exitCode=0 Dec 10 14:38:44 crc kubenswrapper[4847]: I1210 14:38:44.883010 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerDied","Data":"9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd"} Dec 10 14:38:46 crc kubenswrapper[4847]: I1210 14:38:46.896235 4847 generic.go:334] "Generic (PLEG): container finished" podID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerID="50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3" exitCode=0 Dec 10 14:38:46 crc kubenswrapper[4847]: I1210 14:38:46.896277 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerDied","Data":"50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3"} Dec 10 14:38:46 crc kubenswrapper[4847]: I1210 14:38:46.974596 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:46 crc kubenswrapper[4847]: I1210 14:38:46.975458 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:47 crc kubenswrapper[4847]: I1210 14:38:47.012856 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:47 crc kubenswrapper[4847]: I1210 14:38:47.903350 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerStarted","Data":"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983"} Dec 10 14:38:47 crc kubenswrapper[4847]: I1210 14:38:47.921409 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jbhg8" podStartSLOduration=2.468395097 podStartE2EDuration="4.921390252s" podCreationTimestamp="2025-12-10 14:38:43 +0000 UTC" firstStartedPulling="2025-12-10 14:38:44.885069002 +0000 UTC m=+874.454286642" lastFinishedPulling="2025-12-10 14:38:47.338064167 +0000 UTC m=+876.907281797" observedRunningTime="2025-12-10 14:38:47.919422937 +0000 UTC m=+877.488640577" watchObservedRunningTime="2025-12-10 14:38:47.921390252 +0000 UTC m=+877.490607882" Dec 10 14:38:47 crc kubenswrapper[4847]: I1210 14:38:47.930985 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-ptjmg" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.486617 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq"] Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.488151 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.490061 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-dgdxs" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.498763 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.498834 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjlhm\" (UniqueName: \"kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.499201 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.500683 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq"] Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.600126 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.600178 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjlhm\" (UniqueName: \"kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.600277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.601123 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.601125 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.622240 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjlhm\" (UniqueName: \"kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm\") pod \"5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:49 crc kubenswrapper[4847]: I1210 14:38:49.810127 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:50 crc kubenswrapper[4847]: I1210 14:38:50.254104 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq"] Dec 10 14:38:50 crc kubenswrapper[4847]: I1210 14:38:50.920570 4847 generic.go:334] "Generic (PLEG): container finished" podID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerID="40d7049275ec35ebbc87df511bc9a6b63c30bfd6ba7c9ed70395984d1d44b399" exitCode=0 Dec 10 14:38:50 crc kubenswrapper[4847]: I1210 14:38:50.921782 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" event={"ID":"f5631548-ea6d-4a31-a0ae-622f343829fa","Type":"ContainerDied","Data":"40d7049275ec35ebbc87df511bc9a6b63c30bfd6ba7c9ed70395984d1d44b399"} Dec 10 14:38:50 crc kubenswrapper[4847]: I1210 14:38:50.921917 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" event={"ID":"f5631548-ea6d-4a31-a0ae-622f343829fa","Type":"ContainerStarted","Data":"4638510a0acc231566b6d77df6a809e17c10fd5f24312f4794e71ed146d73f38"} Dec 10 14:38:51 crc kubenswrapper[4847]: I1210 14:38:51.930773 4847 generic.go:334] "Generic (PLEG): container finished" podID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerID="dd3a819068cc03b77498de25420703a93d89728d32c0f5f96472c6cfeeb4fcbb" exitCode=0 Dec 10 14:38:51 crc kubenswrapper[4847]: I1210 14:38:51.930924 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" event={"ID":"f5631548-ea6d-4a31-a0ae-622f343829fa","Type":"ContainerDied","Data":"dd3a819068cc03b77498de25420703a93d89728d32c0f5f96472c6cfeeb4fcbb"} Dec 10 14:38:52 crc kubenswrapper[4847]: I1210 14:38:52.938651 4847 generic.go:334] "Generic (PLEG): container finished" podID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerID="22b139e5c8c75a241b760a409b59b9723bf7cc92045c02535bfd2282861623c0" exitCode=0 Dec 10 14:38:52 crc kubenswrapper[4847]: I1210 14:38:52.938692 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" event={"ID":"f5631548-ea6d-4a31-a0ae-622f343829fa","Type":"ContainerDied","Data":"22b139e5c8c75a241b760a409b59b9723bf7cc92045c02535bfd2282861623c0"} Dec 10 14:38:53 crc kubenswrapper[4847]: I1210 14:38:53.589640 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:53 crc kubenswrapper[4847]: I1210 14:38:53.589702 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:53 crc kubenswrapper[4847]: I1210 14:38:53.631443 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.024366 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.245986 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.385776 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util\") pod \"f5631548-ea6d-4a31-a0ae-622f343829fa\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.385900 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjlhm\" (UniqueName: \"kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm\") pod \"f5631548-ea6d-4a31-a0ae-622f343829fa\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.386003 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle\") pod \"f5631548-ea6d-4a31-a0ae-622f343829fa\" (UID: \"f5631548-ea6d-4a31-a0ae-622f343829fa\") " Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.386595 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle" (OuterVolumeSpecName: "bundle") pod "f5631548-ea6d-4a31-a0ae-622f343829fa" (UID: "f5631548-ea6d-4a31-a0ae-622f343829fa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.386826 4847 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.392306 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm" (OuterVolumeSpecName: "kube-api-access-rjlhm") pod "f5631548-ea6d-4a31-a0ae-622f343829fa" (UID: "f5631548-ea6d-4a31-a0ae-622f343829fa"). InnerVolumeSpecName "kube-api-access-rjlhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.399577 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util" (OuterVolumeSpecName: "util") pod "f5631548-ea6d-4a31-a0ae-622f343829fa" (UID: "f5631548-ea6d-4a31-a0ae-622f343829fa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.488149 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjlhm\" (UniqueName: \"kubernetes.io/projected/f5631548-ea6d-4a31-a0ae-622f343829fa-kube-api-access-rjlhm\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.488197 4847 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5631548-ea6d-4a31-a0ae-622f343829fa-util\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.994774 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" event={"ID":"f5631548-ea6d-4a31-a0ae-622f343829fa","Type":"ContainerDied","Data":"4638510a0acc231566b6d77df6a809e17c10fd5f24312f4794e71ed146d73f38"} Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.994806 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4638510a0acc231566b6d77df6a809e17c10fd5f24312f4794e71ed146d73f38" Dec 10 14:38:54 crc kubenswrapper[4847]: I1210 14:38:54.995212 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq" Dec 10 14:38:55 crc kubenswrapper[4847]: I1210 14:38:55.843327 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.001213 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jbhg8" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="registry-server" containerID="cri-o://1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983" gracePeriod=2 Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.338449 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.513198 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2vbl\" (UniqueName: \"kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl\") pod \"27d16898-1cf0-46c0-869c-863f4a4bd134\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.513255 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content\") pod \"27d16898-1cf0-46c0-869c-863f4a4bd134\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.513295 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities\") pod \"27d16898-1cf0-46c0-869c-863f4a4bd134\" (UID: \"27d16898-1cf0-46c0-869c-863f4a4bd134\") " Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.514423 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities" (OuterVolumeSpecName: "utilities") pod "27d16898-1cf0-46c0-869c-863f4a4bd134" (UID: "27d16898-1cf0-46c0-869c-863f4a4bd134"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.520262 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl" (OuterVolumeSpecName: "kube-api-access-v2vbl") pod "27d16898-1cf0-46c0-869c-863f4a4bd134" (UID: "27d16898-1cf0-46c0-869c-863f4a4bd134"). InnerVolumeSpecName "kube-api-access-v2vbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.536543 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27d16898-1cf0-46c0-869c-863f4a4bd134" (UID: "27d16898-1cf0-46c0-869c-863f4a4bd134"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.614951 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.615001 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27d16898-1cf0-46c0-869c-863f4a4bd134-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:56 crc kubenswrapper[4847]: I1210 14:38:56.615015 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2vbl\" (UniqueName: \"kubernetes.io/projected/27d16898-1cf0-46c0-869c-863f4a4bd134-kube-api-access-v2vbl\") on node \"crc\" DevicePath \"\"" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.013094 4847 generic.go:334] "Generic (PLEG): container finished" podID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerID="1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983" exitCode=0 Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.013150 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerDied","Data":"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983"} Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.013189 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbhg8" event={"ID":"27d16898-1cf0-46c0-869c-863f4a4bd134","Type":"ContainerDied","Data":"438ff6e83ffa7b06912098160d04654e41df6f4c049b09ab281829398690a93d"} Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.013211 4847 scope.go:117] "RemoveContainer" containerID="1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.013223 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbhg8" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.037689 4847 scope.go:117] "RemoveContainer" containerID="50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.038990 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.044775 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbhg8"] Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.066122 4847 scope.go:117] "RemoveContainer" containerID="9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.083913 4847 scope.go:117] "RemoveContainer" containerID="1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983" Dec 10 14:38:57 crc kubenswrapper[4847]: E1210 14:38:57.084346 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983\": container with ID starting with 1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983 not found: ID does not exist" containerID="1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.084379 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983"} err="failed to get container status \"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983\": rpc error: code = NotFound desc = could not find container \"1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983\": container with ID starting with 1a67f25b2fb146f1efd77e337172a7adbcd788f2dec54a66614e828773a76983 not found: ID does not exist" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.084399 4847 scope.go:117] "RemoveContainer" containerID="50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3" Dec 10 14:38:57 crc kubenswrapper[4847]: E1210 14:38:57.084673 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3\": container with ID starting with 50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3 not found: ID does not exist" containerID="50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.084699 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3"} err="failed to get container status \"50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3\": rpc error: code = NotFound desc = could not find container \"50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3\": container with ID starting with 50ba50e11b5bc5f571b0ca5350fbc41598a9942dbf8e23f3689b7430068686d3 not found: ID does not exist" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.084730 4847 scope.go:117] "RemoveContainer" containerID="9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd" Dec 10 14:38:57 crc kubenswrapper[4847]: E1210 14:38:57.085051 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd\": container with ID starting with 9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd not found: ID does not exist" containerID="9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd" Dec 10 14:38:57 crc kubenswrapper[4847]: I1210 14:38:57.085073 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd"} err="failed to get container status \"9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd\": rpc error: code = NotFound desc = could not find container \"9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd\": container with ID starting with 9aef5972a3f8181d4d8daf14f6beccb0957adb8b169a9a700ee68bd190e6cbbd not found: ID does not exist" Dec 10 14:38:58 crc kubenswrapper[4847]: I1210 14:38:58.766274 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" path="/var/lib/kubelet/pods/27d16898-1cf0-46c0-869c-863f4a4bd134/volumes" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346013 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh"] Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346298 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="util" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346318 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="util" Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346332 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="extract" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346342 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="extract" Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346353 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="extract-utilities" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346361 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="extract-utilities" Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346376 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="pull" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346383 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="pull" Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346398 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="registry-server" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346405 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="registry-server" Dec 10 14:38:59 crc kubenswrapper[4847]: E1210 14:38:59.346418 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="extract-content" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346426 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="extract-content" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346546 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5631548-ea6d-4a31-a0ae-622f343829fa" containerName="extract" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.346565 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d16898-1cf0-46c0-869c-863f4a4bd134" containerName="registry-server" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.347057 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.348494 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-tqswh" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.350325 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7q5d\" (UniqueName: \"kubernetes.io/projected/caaff4fe-5b58-4a77-b3fe-217742051274-kube-api-access-m7q5d\") pod \"openstack-operator-controller-operator-6646c79bbd-rqsnh\" (UID: \"caaff4fe-5b58-4a77-b3fe-217742051274\") " pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.369171 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh"] Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.451859 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7q5d\" (UniqueName: \"kubernetes.io/projected/caaff4fe-5b58-4a77-b3fe-217742051274-kube-api-access-m7q5d\") pod \"openstack-operator-controller-operator-6646c79bbd-rqsnh\" (UID: \"caaff4fe-5b58-4a77-b3fe-217742051274\") " pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.472316 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7q5d\" (UniqueName: \"kubernetes.io/projected/caaff4fe-5b58-4a77-b3fe-217742051274-kube-api-access-m7q5d\") pod \"openstack-operator-controller-operator-6646c79bbd-rqsnh\" (UID: \"caaff4fe-5b58-4a77-b3fe-217742051274\") " pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:38:59 crc kubenswrapper[4847]: I1210 14:38:59.666909 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:39:00 crc kubenswrapper[4847]: I1210 14:39:00.074849 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh"] Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.011231 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.011780 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.040501 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" event={"ID":"caaff4fe-5b58-4a77-b3fe-217742051274","Type":"ContainerStarted","Data":"7b4db39d3c936a3f43642f2757d0cfde949891c0fa62248fa030095df60aa85c"} Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.251172 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.255725 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.272357 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.378820 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4k27\" (UniqueName: \"kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.378892 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.378979 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.480543 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.481032 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.481099 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4k27\" (UniqueName: \"kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.481872 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.483371 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.500962 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4k27\" (UniqueName: \"kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27\") pod \"community-operators-njd9j\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:01 crc kubenswrapper[4847]: I1210 14:39:01.597484 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:05 crc kubenswrapper[4847]: I1210 14:39:05.025781 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:05 crc kubenswrapper[4847]: W1210 14:39:05.037963 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaa1bc3f_cbec_44bd_a8fe_38b97665bac9.slice/crio-178b984fadc2e201858d045e78e66cb1f9ebb0b8da3cb7678d1978145b07cfca WatchSource:0}: Error finding container 178b984fadc2e201858d045e78e66cb1f9ebb0b8da3cb7678d1978145b07cfca: Status 404 returned error can't find the container with id 178b984fadc2e201858d045e78e66cb1f9ebb0b8da3cb7678d1978145b07cfca Dec 10 14:39:05 crc kubenswrapper[4847]: I1210 14:39:05.063662 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" event={"ID":"caaff4fe-5b58-4a77-b3fe-217742051274","Type":"ContainerStarted","Data":"ac6afb67e49a0169b6ea69cafebd43059ea9aa89b5880400cc0a528f3900218a"} Dec 10 14:39:05 crc kubenswrapper[4847]: I1210 14:39:05.063778 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:39:05 crc kubenswrapper[4847]: I1210 14:39:05.064620 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerStarted","Data":"178b984fadc2e201858d045e78e66cb1f9ebb0b8da3cb7678d1978145b07cfca"} Dec 10 14:39:05 crc kubenswrapper[4847]: I1210 14:39:05.093859 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" podStartSLOduration=1.471525222 podStartE2EDuration="6.093842184s" podCreationTimestamp="2025-12-10 14:38:59 +0000 UTC" firstStartedPulling="2025-12-10 14:39:00.078071608 +0000 UTC m=+889.647289238" lastFinishedPulling="2025-12-10 14:39:04.70038857 +0000 UTC m=+894.269606200" observedRunningTime="2025-12-10 14:39:05.089214865 +0000 UTC m=+894.658432515" watchObservedRunningTime="2025-12-10 14:39:05.093842184 +0000 UTC m=+894.663059814" Dec 10 14:39:06 crc kubenswrapper[4847]: I1210 14:39:06.072869 4847 generic.go:334] "Generic (PLEG): container finished" podID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerID="07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22" exitCode=0 Dec 10 14:39:06 crc kubenswrapper[4847]: I1210 14:39:06.072970 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerDied","Data":"07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22"} Dec 10 14:39:08 crc kubenswrapper[4847]: I1210 14:39:08.087055 4847 generic.go:334] "Generic (PLEG): container finished" podID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerID="92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f" exitCode=0 Dec 10 14:39:08 crc kubenswrapper[4847]: I1210 14:39:08.087143 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerDied","Data":"92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f"} Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.652258 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.654036 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.659896 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.670162 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6646c79bbd-rqsnh" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.801699 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.801831 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.801856 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zggn5\" (UniqueName: \"kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.903321 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.903372 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zggn5\" (UniqueName: \"kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.903472 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.904275 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.904422 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:09 crc kubenswrapper[4847]: I1210 14:39:09.922772 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zggn5\" (UniqueName: \"kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5\") pod \"certified-operators-zcsbl\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:10 crc kubenswrapper[4847]: I1210 14:39:10.048591 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:10 crc kubenswrapper[4847]: I1210 14:39:10.113365 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerStarted","Data":"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1"} Dec 10 14:39:10 crc kubenswrapper[4847]: I1210 14:39:10.135919 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-njd9j" podStartSLOduration=5.549352135 podStartE2EDuration="9.135897539s" podCreationTimestamp="2025-12-10 14:39:01 +0000 UTC" firstStartedPulling="2025-12-10 14:39:06.074221251 +0000 UTC m=+895.643438881" lastFinishedPulling="2025-12-10 14:39:09.660766635 +0000 UTC m=+899.229984285" observedRunningTime="2025-12-10 14:39:10.134552852 +0000 UTC m=+899.703770492" watchObservedRunningTime="2025-12-10 14:39:10.135897539 +0000 UTC m=+899.705115169" Dec 10 14:39:10 crc kubenswrapper[4847]: I1210 14:39:10.340381 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:11 crc kubenswrapper[4847]: I1210 14:39:11.120438 4847 generic.go:334] "Generic (PLEG): container finished" podID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerID="c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64" exitCode=0 Dec 10 14:39:11 crc kubenswrapper[4847]: I1210 14:39:11.120646 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerDied","Data":"c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64"} Dec 10 14:39:11 crc kubenswrapper[4847]: I1210 14:39:11.121237 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerStarted","Data":"3dccb376e755ec15ccaa4f73827859fea21934d6e4ad641682f6ae2a6cd631a6"} Dec 10 14:39:11 crc kubenswrapper[4847]: I1210 14:39:11.598152 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:11 crc kubenswrapper[4847]: I1210 14:39:11.598631 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:12 crc kubenswrapper[4847]: I1210 14:39:12.634889 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-njd9j" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="registry-server" probeResult="failure" output=< Dec 10 14:39:12 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:39:12 crc kubenswrapper[4847]: > Dec 10 14:39:13 crc kubenswrapper[4847]: I1210 14:39:13.133202 4847 generic.go:334] "Generic (PLEG): container finished" podID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerID="803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0" exitCode=0 Dec 10 14:39:13 crc kubenswrapper[4847]: I1210 14:39:13.133312 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerDied","Data":"803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0"} Dec 10 14:39:14 crc kubenswrapper[4847]: I1210 14:39:14.157380 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerStarted","Data":"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2"} Dec 10 14:39:14 crc kubenswrapper[4847]: I1210 14:39:14.174652 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zcsbl" podStartSLOduration=2.718093925 podStartE2EDuration="5.174632659s" podCreationTimestamp="2025-12-10 14:39:09 +0000 UTC" firstStartedPulling="2025-12-10 14:39:11.122000876 +0000 UTC m=+900.691218506" lastFinishedPulling="2025-12-10 14:39:13.57853961 +0000 UTC m=+903.147757240" observedRunningTime="2025-12-10 14:39:14.173506388 +0000 UTC m=+903.742724038" watchObservedRunningTime="2025-12-10 14:39:14.174632659 +0000 UTC m=+903.743850289" Dec 10 14:39:20 crc kubenswrapper[4847]: I1210 14:39:20.048964 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:20 crc kubenswrapper[4847]: I1210 14:39:20.049581 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:20 crc kubenswrapper[4847]: I1210 14:39:20.096208 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:20 crc kubenswrapper[4847]: I1210 14:39:20.235838 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:21 crc kubenswrapper[4847]: I1210 14:39:21.642818 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:21 crc kubenswrapper[4847]: I1210 14:39:21.699853 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:22 crc kubenswrapper[4847]: I1210 14:39:22.645372 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:22 crc kubenswrapper[4847]: I1210 14:39:22.645694 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zcsbl" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="registry-server" containerID="cri-o://84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2" gracePeriod=2 Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.076432 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.182783 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zggn5\" (UniqueName: \"kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5\") pod \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.182902 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities\") pod \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.182948 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content\") pod \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\" (UID: \"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76\") " Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.186248 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities" (OuterVolumeSpecName: "utilities") pod "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" (UID: "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.192916 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5" (OuterVolumeSpecName: "kube-api-access-zggn5") pod "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" (UID: "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76"). InnerVolumeSpecName "kube-api-access-zggn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.217200 4847 generic.go:334] "Generic (PLEG): container finished" podID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerID="84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2" exitCode=0 Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.217287 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerDied","Data":"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2"} Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.217358 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcsbl" event={"ID":"7e2d8d4a-f6b1-43b8-a424-a726e2e65a76","Type":"ContainerDied","Data":"3dccb376e755ec15ccaa4f73827859fea21934d6e4ad641682f6ae2a6cd631a6"} Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.217382 4847 scope.go:117] "RemoveContainer" containerID="84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.217627 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcsbl" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.246388 4847 scope.go:117] "RemoveContainer" containerID="803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.271822 4847 scope.go:117] "RemoveContainer" containerID="c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.289228 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zggn5\" (UniqueName: \"kubernetes.io/projected/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-kube-api-access-zggn5\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.289266 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.289398 4847 scope.go:117] "RemoveContainer" containerID="84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2" Dec 10 14:39:23 crc kubenswrapper[4847]: E1210 14:39:23.289874 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2\": container with ID starting with 84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2 not found: ID does not exist" containerID="84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.289914 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2"} err="failed to get container status \"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2\": rpc error: code = NotFound desc = could not find container \"84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2\": container with ID starting with 84bf266d0d8c0553ed2fc8a0d190a459926b0d28ff4dd17481990a335649b0a2 not found: ID does not exist" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.289939 4847 scope.go:117] "RemoveContainer" containerID="803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0" Dec 10 14:39:23 crc kubenswrapper[4847]: E1210 14:39:23.290153 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0\": container with ID starting with 803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0 not found: ID does not exist" containerID="803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.290173 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0"} err="failed to get container status \"803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0\": rpc error: code = NotFound desc = could not find container \"803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0\": container with ID starting with 803d73eee5213ca29863df87b387096a61c57897751f5019401967f9273a9ce0 not found: ID does not exist" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.290186 4847 scope.go:117] "RemoveContainer" containerID="c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64" Dec 10 14:39:23 crc kubenswrapper[4847]: E1210 14:39:23.290343 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64\": container with ID starting with c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64 not found: ID does not exist" containerID="c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.290364 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64"} err="failed to get container status \"c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64\": rpc error: code = NotFound desc = could not find container \"c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64\": container with ID starting with c5b366c533c0c46555263c67a2606a479e3e4d4b0809fa9a3537b792e1c75b64 not found: ID does not exist" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.626908 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" (UID: "7e2d8d4a-f6b1-43b8-a424-a726e2e65a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.694702 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.860623 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:23 crc kubenswrapper[4847]: I1210 14:39:23.872664 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zcsbl"] Dec 10 14:39:24 crc kubenswrapper[4847]: I1210 14:39:24.767669 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" path="/var/lib/kubelet/pods/7e2d8d4a-f6b1-43b8-a424-a726e2e65a76/volumes" Dec 10 14:39:25 crc kubenswrapper[4847]: I1210 14:39:25.240405 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:25 crc kubenswrapper[4847]: I1210 14:39:25.240639 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-njd9j" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="registry-server" containerID="cri-o://a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1" gracePeriod=2 Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.101659 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.225794 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4k27\" (UniqueName: \"kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27\") pod \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.225842 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities\") pod \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.225913 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content\") pod \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\" (UID: \"caa1bc3f-cbec-44bd-a8fe-38b97665bac9\") " Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.226632 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities" (OuterVolumeSpecName: "utilities") pod "caa1bc3f-cbec-44bd-a8fe-38b97665bac9" (UID: "caa1bc3f-cbec-44bd-a8fe-38b97665bac9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.241060 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27" (OuterVolumeSpecName: "kube-api-access-c4k27") pod "caa1bc3f-cbec-44bd-a8fe-38b97665bac9" (UID: "caa1bc3f-cbec-44bd-a8fe-38b97665bac9"). InnerVolumeSpecName "kube-api-access-c4k27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.247802 4847 generic.go:334] "Generic (PLEG): container finished" podID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerID="a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1" exitCode=0 Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.247849 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerDied","Data":"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1"} Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.247876 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-njd9j" event={"ID":"caa1bc3f-cbec-44bd-a8fe-38b97665bac9","Type":"ContainerDied","Data":"178b984fadc2e201858d045e78e66cb1f9ebb0b8da3cb7678d1978145b07cfca"} Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.247892 4847 scope.go:117] "RemoveContainer" containerID="a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.248112 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-njd9j" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.267798 4847 scope.go:117] "RemoveContainer" containerID="92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.284485 4847 scope.go:117] "RemoveContainer" containerID="07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.288085 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caa1bc3f-cbec-44bd-a8fe-38b97665bac9" (UID: "caa1bc3f-cbec-44bd-a8fe-38b97665bac9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.306545 4847 scope.go:117] "RemoveContainer" containerID="a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1" Dec 10 14:39:26 crc kubenswrapper[4847]: E1210 14:39:26.312134 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1\": container with ID starting with a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1 not found: ID does not exist" containerID="a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.312176 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1"} err="failed to get container status \"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1\": rpc error: code = NotFound desc = could not find container \"a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1\": container with ID starting with a12bc86ead381188d4f71cd423e903040565632b057096a1cd0cb916a5377ee1 not found: ID does not exist" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.312201 4847 scope.go:117] "RemoveContainer" containerID="92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f" Dec 10 14:39:26 crc kubenswrapper[4847]: E1210 14:39:26.312540 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f\": container with ID starting with 92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f not found: ID does not exist" containerID="92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.312625 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f"} err="failed to get container status \"92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f\": rpc error: code = NotFound desc = could not find container \"92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f\": container with ID starting with 92c90d8a928655a226601ea3674d9e0aec216fa469f2d0d526fa0c544ae0712f not found: ID does not exist" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.312693 4847 scope.go:117] "RemoveContainer" containerID="07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22" Dec 10 14:39:26 crc kubenswrapper[4847]: E1210 14:39:26.313072 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22\": container with ID starting with 07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22 not found: ID does not exist" containerID="07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.313099 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22"} err="failed to get container status \"07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22\": rpc error: code = NotFound desc = could not find container \"07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22\": container with ID starting with 07d94c2c2fa519084374ab48781a0f0d372845604b745a24a2765ec1c62c8b22 not found: ID does not exist" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.326985 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.327238 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4k27\" (UniqueName: \"kubernetes.io/projected/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-kube-api-access-c4k27\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.327357 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa1bc3f-cbec-44bd-a8fe-38b97665bac9-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.573203 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.578254 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-njd9j"] Dec 10 14:39:26 crc kubenswrapper[4847]: I1210 14:39:26.766640 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" path="/var/lib/kubelet/pods/caa1bc3f-cbec-44bd-a8fe-38b97665bac9/volumes" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055564 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj"] Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055870 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="extract-content" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055886 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="extract-content" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055901 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="extract-content" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055909 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="extract-content" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055917 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055925 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055938 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055945 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055964 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="extract-utilities" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055972 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="extract-utilities" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.055985 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="extract-utilities" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.055992 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="extract-utilities" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.056117 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2d8d4a-f6b1-43b8-a424-a726e2e65a76" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.056446 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa1bc3f-cbec-44bd-a8fe-38b97665bac9" containerName="registry-server" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.057216 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.059351 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6z9s8" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.065291 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.066355 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.070122 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-h6s82" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.083642 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.085318 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.091239 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-7zkbm" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.102061 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.103060 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.105779 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-n54n5" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.107394 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.117142 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.146884 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.168183 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gvfx\" (UniqueName: \"kubernetes.io/projected/80cfb202-f181-4c1a-b9d6-a7cd036b93db-kube-api-access-2gvfx\") pod \"barbican-operator-controller-manager-7d9dfd778-rqjxj\" (UID: \"80cfb202-f181-4c1a-b9d6-a7cd036b93db\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.193776 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.198773 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.199787 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.203670 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.204435 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kj6wb" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.204918 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.211439 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.215696 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.216110 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7sgfv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.220600 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.221946 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.224021 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-pprc7" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.225420 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.226537 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.229869 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.236073 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.236284 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-glknp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.268907 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkmq4\" (UniqueName: \"kubernetes.io/projected/df30724c-6ddb-4e68-8f57-31f8bb262678-kube-api-access-lkmq4\") pod \"cinder-operator-controller-manager-6c677c69b-pdp9r\" (UID: \"df30724c-6ddb-4e68-8f57-31f8bb262678\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.268968 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhcw6\" (UniqueName: \"kubernetes.io/projected/519d2b7e-57ab-4f4e-a199-68b3fe9a16bf-kube-api-access-zhcw6\") pod \"glance-operator-controller-manager-5697bb5779-fl8dl\" (UID: \"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.268997 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gvfx\" (UniqueName: \"kubernetes.io/projected/80cfb202-f181-4c1a-b9d6-a7cd036b93db-kube-api-access-2gvfx\") pod \"barbican-operator-controller-manager-7d9dfd778-rqjxj\" (UID: \"80cfb202-f181-4c1a-b9d6-a7cd036b93db\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.269052 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9znn7\" (UniqueName: \"kubernetes.io/projected/3c7abfb3-9360-42df-86b1-127066817fe8-kube-api-access-9znn7\") pod \"designate-operator-controller-manager-697fb699cf-pmnjz\" (UID: \"3c7abfb3-9360-42df-86b1-127066817fe8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.269590 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.286758 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.288034 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.289748 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-r9dzj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.319177 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gvfx\" (UniqueName: \"kubernetes.io/projected/80cfb202-f181-4c1a-b9d6-a7cd036b93db-kube-api-access-2gvfx\") pod \"barbican-operator-controller-manager-7d9dfd778-rqjxj\" (UID: \"80cfb202-f181-4c1a-b9d6-a7cd036b93db\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.322869 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.342398 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.343448 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.352557 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mpld6" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.359366 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.360857 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.366011 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.369989 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-5bcdz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.370333 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhcw6\" (UniqueName: \"kubernetes.io/projected/519d2b7e-57ab-4f4e-a199-68b3fe9a16bf-kube-api-access-zhcw6\") pod \"glance-operator-controller-manager-5697bb5779-fl8dl\" (UID: \"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.370394 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lxl\" (UniqueName: \"kubernetes.io/projected/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-kube-api-access-l4lxl\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371075 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371129 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqqzl\" (UniqueName: \"kubernetes.io/projected/defc8bda-15a6-4432-8079-9606a8fdb2e9-kube-api-access-bqqzl\") pod \"heat-operator-controller-manager-5f64f6f8bb-t2wb8\" (UID: \"defc8bda-15a6-4432-8079-9606a8fdb2e9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371158 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9znn7\" (UniqueName: \"kubernetes.io/projected/3c7abfb3-9360-42df-86b1-127066817fe8-kube-api-access-9znn7\") pod \"designate-operator-controller-manager-697fb699cf-pmnjz\" (UID: \"3c7abfb3-9360-42df-86b1-127066817fe8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371188 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbrx6\" (UniqueName: \"kubernetes.io/projected/3a197914-3847-44b5-aa44-527bef265cc5-kube-api-access-wbrx6\") pod \"ironic-operator-controller-manager-967d97867-4j2fk\" (UID: \"3a197914-3847-44b5-aa44-527bef265cc5\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371225 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6gpl\" (UniqueName: \"kubernetes.io/projected/82fa6e27-3c61-4a46-a51c-1f9e034a870c-kube-api-access-v6gpl\") pod \"horizon-operator-controller-manager-68c6d99b8f-gdhjq\" (UID: \"82fa6e27-3c61-4a46-a51c-1f9e034a870c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.371280 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkmq4\" (UniqueName: \"kubernetes.io/projected/df30724c-6ddb-4e68-8f57-31f8bb262678-kube-api-access-lkmq4\") pod \"cinder-operator-controller-manager-6c677c69b-pdp9r\" (UID: \"df30724c-6ddb-4e68-8f57-31f8bb262678\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.382149 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.390903 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.392295 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.399108 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bvn6b" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.410730 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9znn7\" (UniqueName: \"kubernetes.io/projected/3c7abfb3-9360-42df-86b1-127066817fe8-kube-api-access-9znn7\") pod \"designate-operator-controller-manager-697fb699cf-pmnjz\" (UID: \"3c7abfb3-9360-42df-86b1-127066817fe8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.412095 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkmq4\" (UniqueName: \"kubernetes.io/projected/df30724c-6ddb-4e68-8f57-31f8bb262678-kube-api-access-lkmq4\") pod \"cinder-operator-controller-manager-6c677c69b-pdp9r\" (UID: \"df30724c-6ddb-4e68-8f57-31f8bb262678\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.417043 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhcw6\" (UniqueName: \"kubernetes.io/projected/519d2b7e-57ab-4f4e-a199-68b3fe9a16bf-kube-api-access-zhcw6\") pod \"glance-operator-controller-manager-5697bb5779-fl8dl\" (UID: \"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.418736 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.429989 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.473970 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.476018 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbrx6\" (UniqueName: \"kubernetes.io/projected/3a197914-3847-44b5-aa44-527bef265cc5-kube-api-access-wbrx6\") pod \"ironic-operator-controller-manager-967d97867-4j2fk\" (UID: \"3a197914-3847-44b5-aa44-527bef265cc5\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.476057 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bkb7\" (UniqueName: \"kubernetes.io/projected/2fa5c5a1-e3d6-424c-b52a-d01af15458cf-kube-api-access-9bkb7\") pod \"keystone-operator-controller-manager-7765d96ddf-55zkq\" (UID: \"2fa5c5a1-e3d6-424c-b52a-d01af15458cf\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.476095 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6gpl\" (UniqueName: \"kubernetes.io/projected/82fa6e27-3c61-4a46-a51c-1f9e034a870c-kube-api-access-v6gpl\") pod \"horizon-operator-controller-manager-68c6d99b8f-gdhjq\" (UID: \"82fa6e27-3c61-4a46-a51c-1f9e034a870c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.481839 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkg4\" (UniqueName: \"kubernetes.io/projected/9f72b089-2518-4144-a558-bfb5c490d594-kube-api-access-jnkg4\") pod \"mariadb-operator-controller-manager-79c8c4686c-sptmg\" (UID: \"9f72b089-2518-4144-a558-bfb5c490d594\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.482036 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lxl\" (UniqueName: \"kubernetes.io/projected/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-kube-api-access-l4lxl\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.482174 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbhjw\" (UniqueName: \"kubernetes.io/projected/161aaad4-8b14-4a9c-b793-a7937e078aef-kube-api-access-xbhjw\") pod \"manila-operator-controller-manager-5b5fd79c9c-q8fkp\" (UID: \"161aaad4-8b14-4a9c-b793-a7937e078aef\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.482233 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.482312 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx95p\" (UniqueName: \"kubernetes.io/projected/3654ef6c-9eef-40ec-acd3-c101a817f345-kube-api-access-tx95p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gjtlq\" (UID: \"3654ef6c-9eef-40ec-acd3-c101a817f345\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.482370 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqqzl\" (UniqueName: \"kubernetes.io/projected/defc8bda-15a6-4432-8079-9606a8fdb2e9-kube-api-access-bqqzl\") pod \"heat-operator-controller-manager-5f64f6f8bb-t2wb8\" (UID: \"defc8bda-15a6-4432-8079-9606a8fdb2e9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.483536 4847 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.483613 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert podName:73f1dc78-cf06-4c8d-a57a-314ce95cc1d5 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:29.983586492 +0000 UTC m=+919.552804132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert") pod "infra-operator-controller-manager-78d48bff9d-twm4t" (UID: "73f1dc78-cf06-4c8d-a57a-314ce95cc1d5") : secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.488350 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c96c4"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.509305 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.510697 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.522366 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.524298 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqqzl\" (UniqueName: \"kubernetes.io/projected/defc8bda-15a6-4432-8079-9606a8fdb2e9-kube-api-access-bqqzl\") pod \"heat-operator-controller-manager-5f64f6f8bb-t2wb8\" (UID: \"defc8bda-15a6-4432-8079-9606a8fdb2e9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.526418 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbrx6\" (UniqueName: \"kubernetes.io/projected/3a197914-3847-44b5-aa44-527bef265cc5-kube-api-access-wbrx6\") pod \"ironic-operator-controller-manager-967d97867-4j2fk\" (UID: \"3a197914-3847-44b5-aa44-527bef265cc5\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.526840 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gngmd" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.529888 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-d45q2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.536140 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c96c4"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.536364 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.540105 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lxl\" (UniqueName: \"kubernetes.io/projected/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-kube-api-access-l4lxl\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.547851 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.549407 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6gpl\" (UniqueName: \"kubernetes.io/projected/82fa6e27-3c61-4a46-a51c-1f9e034a870c-kube-api-access-v6gpl\") pod \"horizon-operator-controller-manager-68c6d99b8f-gdhjq\" (UID: \"82fa6e27-3c61-4a46-a51c-1f9e034a870c\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.570993 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.572315 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.574911 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.575390 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nfl29" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.578907 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.580422 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.584184 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vnx4b" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.589677 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bkb7\" (UniqueName: \"kubernetes.io/projected/2fa5c5a1-e3d6-424c-b52a-d01af15458cf-kube-api-access-9bkb7\") pod \"keystone-operator-controller-manager-7765d96ddf-55zkq\" (UID: \"2fa5c5a1-e3d6-424c-b52a-d01af15458cf\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.589781 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnkg4\" (UniqueName: \"kubernetes.io/projected/9f72b089-2518-4144-a558-bfb5c490d594-kube-api-access-jnkg4\") pod \"mariadb-operator-controller-manager-79c8c4686c-sptmg\" (UID: \"9f72b089-2518-4144-a558-bfb5c490d594\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.589837 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbhjw\" (UniqueName: \"kubernetes.io/projected/161aaad4-8b14-4a9c-b793-a7937e078aef-kube-api-access-xbhjw\") pod \"manila-operator-controller-manager-5b5fd79c9c-q8fkp\" (UID: \"161aaad4-8b14-4a9c-b793-a7937e078aef\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.589888 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx95p\" (UniqueName: \"kubernetes.io/projected/3654ef6c-9eef-40ec-acd3-c101a817f345-kube-api-access-tx95p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gjtlq\" (UID: \"3654ef6c-9eef-40ec-acd3-c101a817f345\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.590453 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-t224h"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.591581 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.594980 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-d7b4j" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.595073 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.607563 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.614689 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx95p\" (UniqueName: \"kubernetes.io/projected/3654ef6c-9eef-40ec-acd3-c101a817f345-kube-api-access-tx95p\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gjtlq\" (UID: \"3654ef6c-9eef-40ec-acd3-c101a817f345\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.616517 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnkg4\" (UniqueName: \"kubernetes.io/projected/9f72b089-2518-4144-a558-bfb5c490d594-kube-api-access-jnkg4\") pod \"mariadb-operator-controller-manager-79c8c4686c-sptmg\" (UID: \"9f72b089-2518-4144-a558-bfb5c490d594\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.617359 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bkb7\" (UniqueName: \"kubernetes.io/projected/2fa5c5a1-e3d6-424c-b52a-d01af15458cf-kube-api-access-9bkb7\") pod \"keystone-operator-controller-manager-7765d96ddf-55zkq\" (UID: \"2fa5c5a1-e3d6-424c-b52a-d01af15458cf\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.618338 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbhjw\" (UniqueName: \"kubernetes.io/projected/161aaad4-8b14-4a9c-b793-a7937e078aef-kube-api-access-xbhjw\") pod \"manila-operator-controller-manager-5b5fd79c9c-q8fkp\" (UID: \"161aaad4-8b14-4a9c-b793-a7937e078aef\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.621074 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.622266 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.640028 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fhzmp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.638470 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.647029 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.651995 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.652104 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.662865 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-mdz76" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.667891 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.669845 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.680590 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-t224h"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.686007 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695047 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695099 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cv6\" (UniqueName: \"kubernetes.io/projected/d78cab92-e99d-41fb-a35d-ff63d3ceba23-kube-api-access-f8cv6\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695129 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grhqd\" (UniqueName: \"kubernetes.io/projected/64a44edf-788e-4286-83af-6072b8d8b437-kube-api-access-grhqd\") pod \"ovn-operator-controller-manager-b6456fdb6-sbtv2\" (UID: \"64a44edf-788e-4286-83af-6072b8d8b437\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695176 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq9b8\" (UniqueName: \"kubernetes.io/projected/79090a78-3da6-4b17-a875-fc8c2d534076-kube-api-access-xq9b8\") pod \"nova-operator-controller-manager-697bc559fc-kds4v\" (UID: \"79090a78-3da6-4b17-a875-fc8c2d534076\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695210 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqgp9\" (UniqueName: \"kubernetes.io/projected/db714bfd-5854-4a09-a6d0-acf7e3c539a4-kube-api-access-dqgp9\") pod \"octavia-operator-controller-manager-998648c74-c96c4\" (UID: \"db714bfd-5854-4a09-a6d0-acf7e3c539a4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695266 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzqcr\" (UniqueName: \"kubernetes.io/projected/971c05b8-38ac-411a-bdc8-b40ed70daab5-kube-api-access-xzqcr\") pod \"telemetry-operator-controller-manager-58d5ff84df-z6f9c\" (UID: \"971c05b8-38ac-411a-bdc8-b40ed70daab5\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695303 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm9cm\" (UniqueName: \"kubernetes.io/projected/9b0a2f2c-971b-40cc-8544-9842f7878d62-kube-api-access-xm9cm\") pod \"swift-operator-controller-manager-9d58d64bc-4dfs5\" (UID: \"9b0a2f2c-971b-40cc-8544-9842f7878d62\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.695329 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq5mn\" (UniqueName: \"kubernetes.io/projected/87dac2ce-ec87-4d12-9ac8-0a79bedbfa62-kube-api-access-pq5mn\") pod \"placement-operator-controller-manager-78f8948974-t224h\" (UID: \"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.701258 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.701833 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.703521 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.720332 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8k92s"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.746973 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8k92s"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.747063 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.754590 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-t84tz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.758279 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.759928 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.761578 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.769437 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-bdvpz" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.793512 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797461 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grhqd\" (UniqueName: \"kubernetes.io/projected/64a44edf-788e-4286-83af-6072b8d8b437-kube-api-access-grhqd\") pod \"ovn-operator-controller-manager-b6456fdb6-sbtv2\" (UID: \"64a44edf-788e-4286-83af-6072b8d8b437\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq9b8\" (UniqueName: \"kubernetes.io/projected/79090a78-3da6-4b17-a875-fc8c2d534076-kube-api-access-xq9b8\") pod \"nova-operator-controller-manager-697bc559fc-kds4v\" (UID: \"79090a78-3da6-4b17-a875-fc8c2d534076\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797566 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqgp9\" (UniqueName: \"kubernetes.io/projected/db714bfd-5854-4a09-a6d0-acf7e3c539a4-kube-api-access-dqgp9\") pod \"octavia-operator-controller-manager-998648c74-c96c4\" (UID: \"db714bfd-5854-4a09-a6d0-acf7e3c539a4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797602 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzqcr\" (UniqueName: \"kubernetes.io/projected/971c05b8-38ac-411a-bdc8-b40ed70daab5-kube-api-access-xzqcr\") pod \"telemetry-operator-controller-manager-58d5ff84df-z6f9c\" (UID: \"971c05b8-38ac-411a-bdc8-b40ed70daab5\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797641 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm9cm\" (UniqueName: \"kubernetes.io/projected/9b0a2f2c-971b-40cc-8544-9842f7878d62-kube-api-access-xm9cm\") pod \"swift-operator-controller-manager-9d58d64bc-4dfs5\" (UID: \"9b0a2f2c-971b-40cc-8544-9842f7878d62\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.797671 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq5mn\" (UniqueName: \"kubernetes.io/projected/87dac2ce-ec87-4d12-9ac8-0a79bedbfa62-kube-api-access-pq5mn\") pod \"placement-operator-controller-manager-78f8948974-t224h\" (UID: \"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.801981 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.802053 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cv6\" (UniqueName: \"kubernetes.io/projected/d78cab92-e99d-41fb-a35d-ff63d3ceba23-kube-api-access-f8cv6\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.802422 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:29 crc kubenswrapper[4847]: E1210 14:39:29.802467 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:30.302451911 +0000 UTC m=+919.871669541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.821786 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.830434 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq5mn\" (UniqueName: \"kubernetes.io/projected/87dac2ce-ec87-4d12-9ac8-0a79bedbfa62-kube-api-access-pq5mn\") pod \"placement-operator-controller-manager-78f8948974-t224h\" (UID: \"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.832311 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqgp9\" (UniqueName: \"kubernetes.io/projected/db714bfd-5854-4a09-a6d0-acf7e3c539a4-kube-api-access-dqgp9\") pod \"octavia-operator-controller-manager-998648c74-c96c4\" (UID: \"db714bfd-5854-4a09-a6d0-acf7e3c539a4\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.846464 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.847370 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cv6\" (UniqueName: \"kubernetes.io/projected/d78cab92-e99d-41fb-a35d-ff63d3ceba23-kube-api-access-f8cv6\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.847920 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.848113 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq9b8\" (UniqueName: \"kubernetes.io/projected/79090a78-3da6-4b17-a875-fc8c2d534076-kube-api-access-xq9b8\") pod \"nova-operator-controller-manager-697bc559fc-kds4v\" (UID: \"79090a78-3da6-4b17-a875-fc8c2d534076\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.853806 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.853974 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.854150 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sfk24" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.860569 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzqcr\" (UniqueName: \"kubernetes.io/projected/971c05b8-38ac-411a-bdc8-b40ed70daab5-kube-api-access-xzqcr\") pod \"telemetry-operator-controller-manager-58d5ff84df-z6f9c\" (UID: \"971c05b8-38ac-411a-bdc8-b40ed70daab5\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.860572 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grhqd\" (UniqueName: \"kubernetes.io/projected/64a44edf-788e-4286-83af-6072b8d8b437-kube-api-access-grhqd\") pod \"ovn-operator-controller-manager-b6456fdb6-sbtv2\" (UID: \"64a44edf-788e-4286-83af-6072b8d8b437\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.861244 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm9cm\" (UniqueName: \"kubernetes.io/projected/9b0a2f2c-971b-40cc-8544-9842f7878d62-kube-api-access-xm9cm\") pod \"swift-operator-controller-manager-9d58d64bc-4dfs5\" (UID: \"9b0a2f2c-971b-40cc-8544-9842f7878d62\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.861275 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.880442 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.905545 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.907349 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77xwl\" (UniqueName: \"kubernetes.io/projected/0c5922f3-c544-4993-b978-4a035ec08e1b-kube-api-access-77xwl\") pod \"watcher-operator-controller-manager-75944c9b7-4dhdf\" (UID: \"0c5922f3-c544-4993-b978-4a035ec08e1b\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.907450 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtzv\" (UniqueName: \"kubernetes.io/projected/c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90-kube-api-access-qgtzv\") pod \"test-operator-controller-manager-5854674fcc-8k92s\" (UID: \"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.957491 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv"] Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.966866 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.968439 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.978057 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-4zvxv" Dec 10 14:39:29 crc kubenswrapper[4847]: I1210 14:39:29.987494 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.001150 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.018290 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019690 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019750 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77xwl\" (UniqueName: \"kubernetes.io/projected/0c5922f3-c544-4993-b978-4a035ec08e1b-kube-api-access-77xwl\") pod \"watcher-operator-controller-manager-75944c9b7-4dhdf\" (UID: \"0c5922f3-c544-4993-b978-4a035ec08e1b\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019795 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019840 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsvdn\" (UniqueName: \"kubernetes.io/projected/b1d011f5-5052-431d-9933-314f65930ce4-kube-api-access-jsvdn\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.019889 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtzv\" (UniqueName: \"kubernetes.io/projected/c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90-kube-api-access-qgtzv\") pod \"test-operator-controller-manager-5854674fcc-8k92s\" (UID: \"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.020363 4847 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.020412 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert podName:73f1dc78-cf06-4c8d-a57a-314ce95cc1d5 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:31.020395263 +0000 UTC m=+920.589612893 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert") pod "infra-operator-controller-manager-78d48bff9d-twm4t" (UID: "73f1dc78-cf06-4c8d-a57a-314ce95cc1d5") : secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.035588 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.052615 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.067526 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtzv\" (UniqueName: \"kubernetes.io/projected/c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90-kube-api-access-qgtzv\") pod \"test-operator-controller-manager-5854674fcc-8k92s\" (UID: \"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.069410 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77xwl\" (UniqueName: \"kubernetes.io/projected/0c5922f3-c544-4993-b978-4a035ec08e1b-kube-api-access-77xwl\") pod \"watcher-operator-controller-manager-75944c9b7-4dhdf\" (UID: \"0c5922f3-c544-4993-b978-4a035ec08e1b\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.124699 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz7lk\" (UniqueName: \"kubernetes.io/projected/4706d44c-19de-49a6-b604-dbe911592e48-kube-api-access-tz7lk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-d5mdv\" (UID: \"4706d44c-19de-49a6-b604-dbe911592e48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.124780 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.124844 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsvdn\" (UniqueName: \"kubernetes.io/projected/b1d011f5-5052-431d-9933-314f65930ce4-kube-api-access-jsvdn\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.124868 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.125021 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.125076 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:30.625059523 +0000 UTC m=+920.194277153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.125521 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.125602 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:30.625574557 +0000 UTC m=+920.194792247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.137314 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.153152 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsvdn\" (UniqueName: \"kubernetes.io/projected/b1d011f5-5052-431d-9933-314f65930ce4-kube-api-access-jsvdn\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.162180 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.212681 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.226842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz7lk\" (UniqueName: \"kubernetes.io/projected/4706d44c-19de-49a6-b604-dbe911592e48-kube-api-access-tz7lk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-d5mdv\" (UID: \"4706d44c-19de-49a6-b604-dbe911592e48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.253189 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz7lk\" (UniqueName: \"kubernetes.io/projected/4706d44c-19de-49a6-b604-dbe911592e48-kube-api-access-tz7lk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-d5mdv\" (UID: \"4706d44c-19de-49a6-b604-dbe911592e48\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.277914 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" event={"ID":"80cfb202-f181-4c1a-b9d6-a7cd036b93db","Type":"ContainerStarted","Data":"31e1a7d3adae5eab99c2919bb524d582512782931b906fa9164dba85a4fd3f76"} Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.293458 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.318695 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.330403 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.330611 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.330662 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:31.330647371 +0000 UTC m=+920.899865001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.349332 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.354917 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.420442 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.635203 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.635346 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.635470 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.635516 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:31.63550146 +0000 UTC m=+921.204719090 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.635556 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: E1210 14:39:30.635576 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:31.635569342 +0000 UTC m=+921.204786972 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.823687 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c96c4"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.834630 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r"] Dec 10 14:39:30 crc kubenswrapper[4847]: W1210 14:39:30.869617 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f72b089_2518_4144_a558_bfb5c490d594.slice/crio-8410f0631f6143a9cc2bc3305b1e44717566cd4035affcda1a755dd50122d41d WatchSource:0}: Error finding container 8410f0631f6143a9cc2bc3305b1e44717566cd4035affcda1a755dd50122d41d: Status 404 returned error can't find the container with id 8410f0631f6143a9cc2bc3305b1e44717566cd4035affcda1a755dd50122d41d Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.870027 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg"] Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.952000 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v"] Dec 10 14:39:30 crc kubenswrapper[4847]: W1210 14:39:30.956607 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79090a78_3da6_4b17_a875_fc8c2d534076.slice/crio-a4b5dbbca77aae6eb5be0203a335990d0cc5139f767378df98c0010dbf894b7d WatchSource:0}: Error finding container a4b5dbbca77aae6eb5be0203a335990d0cc5139f767378df98c0010dbf894b7d: Status 404 returned error can't find the container with id a4b5dbbca77aae6eb5be0203a335990d0cc5139f767378df98c0010dbf894b7d Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.958389 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq"] Dec 10 14:39:30 crc kubenswrapper[4847]: W1210 14:39:30.968257 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82fa6e27_3c61_4a46_a51c_1f9e034a870c.slice/crio-ba0b0eb453b029b648846c1ec77ddc3f2998a79c63af66b7608b2e57e3034886 WatchSource:0}: Error finding container ba0b0eb453b029b648846c1ec77ddc3f2998a79c63af66b7608b2e57e3034886: Status 404 returned error can't find the container with id ba0b0eb453b029b648846c1ec77ddc3f2998a79c63af66b7608b2e57e3034886 Dec 10 14:39:30 crc kubenswrapper[4847]: I1210 14:39:30.976676 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq"] Dec 10 14:39:30 crc kubenswrapper[4847]: W1210 14:39:30.980327 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3654ef6c_9eef_40ec_acd3_c101a817f345.slice/crio-ff3db37d776ce8021c62f919433321c67a020981736758be90d068da72be8266 WatchSource:0}: Error finding container ff3db37d776ce8021c62f919433321c67a020981736758be90d068da72be8266: Status 404 returned error can't find the container with id ff3db37d776ce8021c62f919433321c67a020981736758be90d068da72be8266 Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.011063 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.011122 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.049535 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.049839 4847 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.049904 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert podName:73f1dc78-cf06-4c8d-a57a-314ce95cc1d5 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:33.049884816 +0000 UTC m=+922.619102446 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert") pod "infra-operator-controller-manager-78d48bff9d-twm4t" (UID: "73f1dc78-cf06-4c8d-a57a-314ce95cc1d5") : secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.140020 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2"] Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.152417 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-t224h"] Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.202695 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-grhqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-sbtv2_openstack-operators(64a44edf-788e-4286-83af-6072b8d8b437): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.204674 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-grhqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-sbtv2_openstack-operators(64a44edf-788e-4286-83af-6072b8d8b437): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.206026 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" podUID="64a44edf-788e-4286-83af-6072b8d8b437" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.206730 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzqcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-z6f9c_openstack-operators(971c05b8-38ac-411a-bdc8-b40ed70daab5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.206960 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5"] Dec 10 14:39:31 crc kubenswrapper[4847]: W1210 14:39:31.212410 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87dac2ce_ec87_4d12_9ac8_0a79bedbfa62.slice/crio-dd27f739a003819b02361d8924b9d0f3c6b958c418f43adca76b477fc83b822f WatchSource:0}: Error finding container dd27f739a003819b02361d8924b9d0f3c6b958c418f43adca76b477fc83b822f: Status 404 returned error can't find the container with id dd27f739a003819b02361d8924b9d0f3c6b958c418f43adca76b477fc83b822f Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.213607 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzqcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-z6f9c_openstack-operators(971c05b8-38ac-411a-bdc8-b40ed70daab5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.216012 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" podUID="971c05b8-38ac-411a-bdc8-b40ed70daab5" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.217907 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c"] Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.227088 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pq5mn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-t224h_openstack-operators(87dac2ce-ec87-4d12-9ac8-0a79bedbfa62): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.229892 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pq5mn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-t224h_openstack-operators(87dac2ce-ec87-4d12-9ac8-0a79bedbfa62): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.231266 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" podUID="87dac2ce-ec87-4d12-9ac8-0a79bedbfa62" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.248897 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8k92s"] Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.255003 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz"] Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.298197 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" event={"ID":"9b0a2f2c-971b-40cc-8544-9842f7878d62","Type":"ContainerStarted","Data":"abd2f156053284486c68f841bbe0c68fe33ee8b47d5cda96bdcb9adcb999ef30"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.307341 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" event={"ID":"3654ef6c-9eef-40ec-acd3-c101a817f345","Type":"ContainerStarted","Data":"ff3db37d776ce8021c62f919433321c67a020981736758be90d068da72be8266"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.311112 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" event={"ID":"defc8bda-15a6-4432-8079-9606a8fdb2e9","Type":"ContainerStarted","Data":"644b136e6f546081db80b68cbfe427e888defd6e6b7887ae29c7a17b0b3279c7"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.316877 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" event={"ID":"971c05b8-38ac-411a-bdc8-b40ed70daab5","Type":"ContainerStarted","Data":"92f8078f997c56b0bef695e0e9f3326311a1c694ef5c07fc1cc1c11dd42c37a0"} Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.328603 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" podUID="971c05b8-38ac-411a-bdc8-b40ed70daab5" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.330810 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf"] Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.332340 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" event={"ID":"9f72b089-2518-4144-a558-bfb5c490d594","Type":"ContainerStarted","Data":"8410f0631f6143a9cc2bc3305b1e44717566cd4035affcda1a755dd50122d41d"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.334163 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" event={"ID":"161aaad4-8b14-4a9c-b793-a7937e078aef","Type":"ContainerStarted","Data":"50bfcc02369f7692bc30d6780f49ca121cfca301efd26f31fb6eaaa9d304b5ed"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.355056 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.355255 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.355311 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:33.35529251 +0000 UTC m=+922.924510140 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.361086 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" event={"ID":"2fa5c5a1-e3d6-424c-b52a-d01af15458cf","Type":"ContainerStarted","Data":"29c760f90af2a9674eedfdd21287a4004b346c817eb34496655291b660953b36"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.371821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" event={"ID":"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90","Type":"ContainerStarted","Data":"2f2218221acf88640b9b0b504962c479e04a624bbdc08f5e4221916d72ee4d07"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.379029 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" event={"ID":"3c7abfb3-9360-42df-86b1-127066817fe8","Type":"ContainerStarted","Data":"e8e991316d8f72a99a09e5ca607acadbcd1c643199941fb321ff98169dcef167"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.382006 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv"] Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.393651 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" event={"ID":"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf","Type":"ContainerStarted","Data":"efe154fab822ade0b2291713403847296a77311690312845b28c0ec5535b0b57"} Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.399085 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tz7lk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-d5mdv_openstack-operators(4706d44c-19de-49a6-b604-dbe911592e48): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.399859 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" event={"ID":"79090a78-3da6-4b17-a875-fc8c2d534076","Type":"ContainerStarted","Data":"a4b5dbbca77aae6eb5be0203a335990d0cc5139f767378df98c0010dbf894b7d"} Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.400999 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" podUID="4706d44c-19de-49a6-b604-dbe911592e48" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.403962 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" event={"ID":"df30724c-6ddb-4e68-8f57-31f8bb262678","Type":"ContainerStarted","Data":"e980b69d4d08e8b17db08129e8af0a093794cd438ad89a6f06fb040e7da3139c"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.406495 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" event={"ID":"3a197914-3847-44b5-aa44-527bef265cc5","Type":"ContainerStarted","Data":"880ea324362e4a50aac9acdd80cdb337aa65c112181e1a064b7ef6cd9fe6c702"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.414695 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" event={"ID":"82fa6e27-3c61-4a46-a51c-1f9e034a870c","Type":"ContainerStarted","Data":"ba0b0eb453b029b648846c1ec77ddc3f2998a79c63af66b7608b2e57e3034886"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.417592 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" event={"ID":"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62","Type":"ContainerStarted","Data":"dd27f739a003819b02361d8924b9d0f3c6b958c418f43adca76b477fc83b822f"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.419625 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" event={"ID":"64a44edf-788e-4286-83af-6072b8d8b437","Type":"ContainerStarted","Data":"10780d1b0748739cc108787c206a42df09a4fed08fd9fb4de5e7edd1d1b5d0bd"} Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.422770 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" event={"ID":"db714bfd-5854-4a09-a6d0-acf7e3c539a4","Type":"ContainerStarted","Data":"c0997758769bbcbb9aac0292abb690bd392f775fd63f6912800a111d5a08c579"} Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.432183 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" podUID="64a44edf-788e-4286-83af-6072b8d8b437" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.435941 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" podUID="87dac2ce-ec87-4d12-9ac8-0a79bedbfa62" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.658937 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:31 crc kubenswrapper[4847]: I1210 14:39:31.659388 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.659128 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.659508 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:33.659490651 +0000 UTC m=+923.228708281 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.659582 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:31 crc kubenswrapper[4847]: E1210 14:39:31.659632 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:33.659617755 +0000 UTC m=+923.228835385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:32 crc kubenswrapper[4847]: I1210 14:39:32.438209 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" event={"ID":"0c5922f3-c544-4993-b978-4a035ec08e1b","Type":"ContainerStarted","Data":"be9434b3e5502fb80d26c149f297a8ccbb4b85189ff3cbbf6e2806917e20232e"} Dec 10 14:39:32 crc kubenswrapper[4847]: I1210 14:39:32.440603 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" event={"ID":"4706d44c-19de-49a6-b604-dbe911592e48","Type":"ContainerStarted","Data":"7de1f3873daddf0ecfd836db19a6ea10000b4b36d701e922b1f36a942a2e92f2"} Dec 10 14:39:32 crc kubenswrapper[4847]: E1210 14:39:32.447669 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" podUID="4706d44c-19de-49a6-b604-dbe911592e48" Dec 10 14:39:32 crc kubenswrapper[4847]: E1210 14:39:32.447989 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" podUID="971c05b8-38ac-411a-bdc8-b40ed70daab5" Dec 10 14:39:32 crc kubenswrapper[4847]: E1210 14:39:32.449586 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" podUID="87dac2ce-ec87-4d12-9ac8-0a79bedbfa62" Dec 10 14:39:32 crc kubenswrapper[4847]: E1210 14:39:32.451217 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" podUID="64a44edf-788e-4286-83af-6072b8d8b437" Dec 10 14:39:33 crc kubenswrapper[4847]: I1210 14:39:33.083405 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.083923 4847 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.083972 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert podName:73f1dc78-cf06-4c8d-a57a-314ce95cc1d5 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:37.08395734 +0000 UTC m=+926.653174970 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert") pod "infra-operator-controller-manager-78d48bff9d-twm4t" (UID: "73f1dc78-cf06-4c8d-a57a-314ce95cc1d5") : secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: I1210 14:39:33.394387 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.394642 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.394724 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:37.394690892 +0000 UTC m=+926.963908532 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.463009 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" podUID="4706d44c-19de-49a6-b604-dbe911592e48" Dec 10 14:39:33 crc kubenswrapper[4847]: I1210 14:39:33.698842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:33 crc kubenswrapper[4847]: I1210 14:39:33.698900 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.699034 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.699082 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:37.69906763 +0000 UTC m=+927.268285260 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.699553 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:33 crc kubenswrapper[4847]: E1210 14:39:33.699635 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:37.699614065 +0000 UTC m=+927.268831755 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: I1210 14:39:37.156131 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.156806 4847 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.156866 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert podName:73f1dc78-cf06-4c8d-a57a-314ce95cc1d5 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:45.156847303 +0000 UTC m=+934.726064943 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert") pod "infra-operator-controller-manager-78d48bff9d-twm4t" (UID: "73f1dc78-cf06-4c8d-a57a-314ce95cc1d5") : secret "infra-operator-webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: I1210 14:39:37.459539 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.459768 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.459816 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:45.459800629 +0000 UTC m=+935.029018259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: I1210 14:39:37.762564 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.762866 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: I1210 14:39:37.763003 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.763092 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:45.763075394 +0000 UTC m=+935.332293024 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.763394 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:37 crc kubenswrapper[4847]: E1210 14:39:37.763429 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:39:45.763422095 +0000 UTC m=+935.332639725 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:44 crc kubenswrapper[4847]: E1210 14:39:44.262042 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 10 14:39:44 crc kubenswrapper[4847]: E1210 14:39:44.262916 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lkmq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-pdp9r_openstack-operators(df30724c-6ddb-4e68-8f57-31f8bb262678): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:44 crc kubenswrapper[4847]: E1210 14:39:44.850227 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 10 14:39:44 crc kubenswrapper[4847]: E1210 14:39:44.850505 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xm9cm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-4dfs5_openstack-operators(9b0a2f2c-971b-40cc-8544-9842f7878d62): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.199488 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.215284 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73f1dc78-cf06-4c8d-a57a-314ce95cc1d5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-twm4t\" (UID: \"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.468212 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.514606 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.514811 4847 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.514896 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert podName:d78cab92-e99d-41fb-a35d-ff63d3ceba23 nodeName:}" failed. No retries permitted until 2025-12-10 14:40:01.514871778 +0000 UTC m=+951.084089498 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f2swnv" (UID: "d78cab92-e99d-41fb-a35d-ff63d3ceba23") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.638075 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.638272 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qgtzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-8k92s_openstack-operators(c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.817817 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:45 crc kubenswrapper[4847]: I1210 14:39:45.817895 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.818082 4847 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.818136 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:40:01.818119132 +0000 UTC m=+951.387336772 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "metrics-server-cert" not found Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.818642 4847 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 14:39:45 crc kubenswrapper[4847]: E1210 14:39:45.818694 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs podName:b1d011f5-5052-431d-9933-314f65930ce4 nodeName:}" failed. No retries permitted until 2025-12-10 14:40:01.818686138 +0000 UTC m=+951.387903768 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs") pod "openstack-operator-controller-manager-79b6c4fddb-8mzvt" (UID: "b1d011f5-5052-431d-9933-314f65930ce4") : secret "webhook-server-cert" not found Dec 10 14:39:46 crc kubenswrapper[4847]: E1210 14:39:46.279154 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 10 14:39:46 crc kubenswrapper[4847]: E1210 14:39:46.279372 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnkg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-sptmg_openstack-operators(9f72b089-2518-4144-a558-bfb5c490d594): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:46 crc kubenswrapper[4847]: E1210 14:39:46.943532 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 10 14:39:46 crc kubenswrapper[4847]: E1210 14:39:46.943742 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6gpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-gdhjq_openstack-operators(82fa6e27-3c61-4a46-a51c-1f9e034a870c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:47 crc kubenswrapper[4847]: E1210 14:39:47.633263 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 10 14:39:47 crc kubenswrapper[4847]: E1210 14:39:47.634309 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tx95p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-gjtlq_openstack-operators(3654ef6c-9eef-40ec-acd3-c101a817f345): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:48 crc kubenswrapper[4847]: E1210 14:39:48.395597 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a" Dec 10 14:39:48 crc kubenswrapper[4847]: E1210 14:39:48.395794 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-77xwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-4dhdf_openstack-operators(0c5922f3-c544-4993-b978-4a035ec08e1b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:49 crc kubenswrapper[4847]: E1210 14:39:49.207006 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 10 14:39:49 crc kubenswrapper[4847]: E1210 14:39:49.207669 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gvfx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-rqjxj_openstack-operators(80cfb202-f181-4c1a-b9d6-a7cd036b93db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:49 crc kubenswrapper[4847]: E1210 14:39:49.867479 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 10 14:39:49 crc kubenswrapper[4847]: E1210 14:39:49.867650 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9bkb7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-55zkq_openstack-operators(2fa5c5a1-e3d6-424c-b52a-d01af15458cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:39:54 crc kubenswrapper[4847]: I1210 14:39:54.173161 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t"] Dec 10 14:39:54 crc kubenswrapper[4847]: W1210 14:39:54.312409 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73f1dc78_cf06_4c8d_a57a_314ce95cc1d5.slice/crio-dd46671eba7e12d79e1dbfa30a4d5046c51f1118f0991be76723dee707280c77 WatchSource:0}: Error finding container dd46671eba7e12d79e1dbfa30a4d5046c51f1118f0991be76723dee707280c77: Status 404 returned error can't find the container with id dd46671eba7e12d79e1dbfa30a4d5046c51f1118f0991be76723dee707280c77 Dec 10 14:39:54 crc kubenswrapper[4847]: I1210 14:39:54.584406 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" event={"ID":"79090a78-3da6-4b17-a875-fc8c2d534076","Type":"ContainerStarted","Data":"08ad17e04962c66b2d87f347fa2b72c17eb0acbfafdddade7d0687237001568b"} Dec 10 14:39:54 crc kubenswrapper[4847]: I1210 14:39:54.585767 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" event={"ID":"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5","Type":"ContainerStarted","Data":"dd46671eba7e12d79e1dbfa30a4d5046c51f1118f0991be76723dee707280c77"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.601833 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" event={"ID":"971c05b8-38ac-411a-bdc8-b40ed70daab5","Type":"ContainerStarted","Data":"3a4eafe7b7fdefdb219d884ac60823f955cea9edcf5499da8582fc56ad39defc"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.615598 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" event={"ID":"161aaad4-8b14-4a9c-b793-a7937e078aef","Type":"ContainerStarted","Data":"57b4265ff2d5941d6a4a6450a5336630322612a7255e1267f105536caed84e74"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.628804 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" event={"ID":"3a197914-3847-44b5-aa44-527bef265cc5","Type":"ContainerStarted","Data":"2908d620aa56fe6af677b1487c1d076f5835ce4157a6593f6b758259533da620"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.649244 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" event={"ID":"defc8bda-15a6-4432-8079-9606a8fdb2e9","Type":"ContainerStarted","Data":"f66f11fd68069f1a889fb0269a38c2907d1b2ad98a247ec198330ec3764594a8"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.654919 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" event={"ID":"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf","Type":"ContainerStarted","Data":"c7cdab0931ab2d689e6e292903e452eaa6adb846da2cb87c7a243151f3ff2580"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.656184 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" event={"ID":"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62","Type":"ContainerStarted","Data":"8066f5e7ee69ea00bef791a0521663f6f4b726627d912da13eb82944a51b4698"} Dec 10 14:39:55 crc kubenswrapper[4847]: I1210 14:39:55.673330 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" event={"ID":"db714bfd-5854-4a09-a6d0-acf7e3c539a4","Type":"ContainerStarted","Data":"8d8613ae26916a64c5e5da566666e0d5da247f3ac694846a25b25b2815d04334"} Dec 10 14:39:56 crc kubenswrapper[4847]: I1210 14:39:56.682753 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" event={"ID":"64a44edf-788e-4286-83af-6072b8d8b437","Type":"ContainerStarted","Data":"d15a5b411365e78b4ec24386d60c6e5ee06e045bc6d138dd1027319d417bf9a8"} Dec 10 14:39:56 crc kubenswrapper[4847]: I1210 14:39:56.684727 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" event={"ID":"3c7abfb3-9360-42df-86b1-127066817fe8","Type":"ContainerStarted","Data":"2bb60583108064011444baa44dbb398ffc98e618b87576d8dcfb5f09f7714726"} Dec 10 14:39:57 crc kubenswrapper[4847]: I1210 14:39:57.709255 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" event={"ID":"4706d44c-19de-49a6-b604-dbe911592e48","Type":"ContainerStarted","Data":"85d238f2ae5e64922e46cc978fdc698faf42b61d50476b39a0a58cd7d6dec297"} Dec 10 14:39:57 crc kubenswrapper[4847]: I1210 14:39:57.733932 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-d5mdv" podStartSLOduration=5.721824354 podStartE2EDuration="28.733910638s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.398922774 +0000 UTC m=+920.968140404" lastFinishedPulling="2025-12-10 14:39:54.411009058 +0000 UTC m=+943.980226688" observedRunningTime="2025-12-10 14:39:57.726666288 +0000 UTC m=+947.295883918" watchObservedRunningTime="2025-12-10 14:39:57.733910638 +0000 UTC m=+947.303128269" Dec 10 14:40:00 crc kubenswrapper[4847]: E1210 14:40:00.610489 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" podUID="82fa6e27-3c61-4a46-a51c-1f9e034a870c" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.739160 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" event={"ID":"3a197914-3847-44b5-aa44-527bef265cc5","Type":"ContainerStarted","Data":"a80f50e559d06213282b9e0a75d88d324ba3c2323108bc0ad91c58521a977338"} Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.739583 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.742112 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" event={"ID":"82fa6e27-3c61-4a46-a51c-1f9e034a870c","Type":"ContainerStarted","Data":"a6dce914926f8ac26c33c2e572df594d6826eb65ac16823af474da77fbd8a632"} Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.746831 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.747853 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" event={"ID":"64a44edf-788e-4286-83af-6072b8d8b437","Type":"ContainerStarted","Data":"fafb5d8eb0f6e71b43a8f12dca0322c5ac5888462571ca877534f4ac057dd5af"} Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.748310 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.753195 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" event={"ID":"3c7abfb3-9360-42df-86b1-127066817fe8","Type":"ContainerStarted","Data":"3065af086b188478f642eb2403a88c51b0942197115cebc279a3c7ffa0746fc7"} Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.753336 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.756028 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" event={"ID":"db714bfd-5854-4a09-a6d0-acf7e3c539a4","Type":"ContainerStarted","Data":"939b283abded6cccc9eacf69bf3223dcf9bd9df43ee5494e82ba4daa96baa85e"} Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.757775 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.772917 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4j2fk" podStartSLOduration=2.819720888 podStartE2EDuration="31.772897073s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.629692329 +0000 UTC m=+920.198909959" lastFinishedPulling="2025-12-10 14:39:59.582868514 +0000 UTC m=+949.152086144" observedRunningTime="2025-12-10 14:40:00.762495203 +0000 UTC m=+950.331712843" watchObservedRunningTime="2025-12-10 14:40:00.772897073 +0000 UTC m=+950.342114703" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.792749 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.813108 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c96c4" podStartSLOduration=3.02334702 podStartE2EDuration="31.8130837s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.863268485 +0000 UTC m=+920.432486115" lastFinishedPulling="2025-12-10 14:39:59.653005165 +0000 UTC m=+949.222222795" observedRunningTime="2025-12-10 14:40:00.800880491 +0000 UTC m=+950.370098121" watchObservedRunningTime="2025-12-10 14:40:00.8130837 +0000 UTC m=+950.382301330" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.834386 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" podStartSLOduration=3.455278164 podStartE2EDuration="31.834366413s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.201590895 +0000 UTC m=+920.770808525" lastFinishedPulling="2025-12-10 14:39:59.580679144 +0000 UTC m=+949.149896774" observedRunningTime="2025-12-10 14:40:00.82097111 +0000 UTC m=+950.390188750" watchObservedRunningTime="2025-12-10 14:40:00.834366413 +0000 UTC m=+950.403584043" Dec 10 14:40:00 crc kubenswrapper[4847]: I1210 14:40:00.866601 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" podStartSLOduration=3.37347049 podStartE2EDuration="31.866560059s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.202318805 +0000 UTC m=+920.771536425" lastFinishedPulling="2025-12-10 14:39:59.695408364 +0000 UTC m=+949.264625994" observedRunningTime="2025-12-10 14:40:00.859419569 +0000 UTC m=+950.428637199" watchObservedRunningTime="2025-12-10 14:40:00.866560059 +0000 UTC m=+950.435777689" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.010756 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.010812 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.010866 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.011453 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.011509 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006" gracePeriod=600 Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.030694 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" podUID="df30724c-6ddb-4e68-8f57-31f8bb262678" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.194562 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" podUID="0c5922f3-c544-4993-b978-4a035ec08e1b" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.347177 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" podUID="3654ef6c-9eef-40ec-acd3-c101a817f345" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.422023 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" podUID="2fa5c5a1-e3d6-424c-b52a-d01af15458cf" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.528486 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.536474 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d78cab92-e99d-41fb-a35d-ff63d3ceba23-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f2swnv\" (UID: \"d78cab92-e99d-41fb-a35d-ff63d3ceba23\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.574273 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" podUID="c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.591111 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" podUID="9f72b089-2518-4144-a558-bfb5c490d594" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.591502 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" podUID="9b0a2f2c-971b-40cc-8544-9842f7878d62" Dec 10 14:40:01 crc kubenswrapper[4847]: E1210 14:40:01.601885 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" podUID="80cfb202-f181-4c1a-b9d6-a7cd036b93db" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.718299 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.832849 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.833171 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.839036 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" event={"ID":"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5","Type":"ContainerStarted","Data":"512dcb315eb3287ceafeb1bbd816002a7a1359497d6d289837cb965bbac8e02c"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.839091 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" event={"ID":"73f1dc78-cf06-4c8d-a57a-314ce95cc1d5","Type":"ContainerStarted","Data":"76b4d97ee1ad16ae2a6e2d746420d15c1dcb92ec012aafb6d296c4e78bee733a"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.840004 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.862619 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-metrics-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.864624 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b1d011f5-5052-431d-9933-314f65930ce4-webhook-certs\") pod \"openstack-operator-controller-manager-79b6c4fddb-8mzvt\" (UID: \"b1d011f5-5052-431d-9933-314f65930ce4\") " pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.873488 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" event={"ID":"80cfb202-f181-4c1a-b9d6-a7cd036b93db","Type":"ContainerStarted","Data":"93b87d40b0457946fd1167de0217fb6cd6fede88d6c968d409cc8bfa0cef7c8c"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.906134 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006" exitCode=0 Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.906237 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.906268 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.906287 4847 scope.go:117] "RemoveContainer" containerID="eff6b5601a8f5384d02b8eb4d030b016e7670663f9d4f00b3f59dfd4f9403fa5" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.910250 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" podStartSLOduration=26.526812594 podStartE2EDuration="32.910232876s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:54.325796729 +0000 UTC m=+943.895014359" lastFinishedPulling="2025-12-10 14:40:00.709217001 +0000 UTC m=+950.278434641" observedRunningTime="2025-12-10 14:40:01.908891799 +0000 UTC m=+951.478109429" watchObservedRunningTime="2025-12-10 14:40:01.910232876 +0000 UTC m=+951.479450516" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.931175 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" event={"ID":"519d2b7e-57ab-4f4e-a199-68b3fe9a16bf","Type":"ContainerStarted","Data":"81f795c45da7e773de56ab9766f990397cc608c6938bb7b15734480f7ddd1671"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.931825 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.944046 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.957984 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" event={"ID":"0c5922f3-c544-4993-b978-4a035ec08e1b","Type":"ContainerStarted","Data":"bb9048e28ae89e850290d9d4c9233948977d87bdb8e65f2d772a21a02f491b3b"} Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.982744 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:01 crc kubenswrapper[4847]: I1210 14:40:01.982831 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" event={"ID":"9b0a2f2c-971b-40cc-8544-9842f7878d62","Type":"ContainerStarted","Data":"7b77422f8194e4feb5d330b8d6ddc992bd36e2a05bf86764fff0785eb7cdbbbf"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.012984 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" event={"ID":"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90","Type":"ContainerStarted","Data":"c9deb7dddce71e1d8dd8a68efdcf615ab100dbfac8e67960190db35593412883"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.047055 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" event={"ID":"87dac2ce-ec87-4d12-9ac8-0a79bedbfa62","Type":"ContainerStarted","Data":"482c294bdcf4e622fe06dca2ba7ed260ce1d0b3977dd8b9e0f8677e1330a7838"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.048399 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.053290 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.073022 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" event={"ID":"9f72b089-2518-4144-a558-bfb5c490d594","Type":"ContainerStarted","Data":"79ca0da16cec6018979b330d914b50ce7fc84b6f155c87c217a83cebba24e579"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.103725 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" event={"ID":"df30724c-6ddb-4e68-8f57-31f8bb262678","Type":"ContainerStarted","Data":"4887b161ea3b80abdc9ce71b931fe9558f8ff6b908f3a7dd8ca4208a0a47cf55"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.130011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" event={"ID":"82fa6e27-3c61-4a46-a51c-1f9e034a870c","Type":"ContainerStarted","Data":"1add6e7fb911055ca22cb950aff0de3f3790c6c5ebdfa5ba44bb75547363e2d0"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.130866 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.180313 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" event={"ID":"3654ef6c-9eef-40ec-acd3-c101a817f345","Type":"ContainerStarted","Data":"5e72b08da85cd63563c5dd0964c1cb4a195c61531b9d2d900978b1d65a099d35"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.186577 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-fl8dl" podStartSLOduration=2.773997216 podStartE2EDuration="33.186557242s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.295385671 +0000 UTC m=+919.864603301" lastFinishedPulling="2025-12-10 14:40:00.707945697 +0000 UTC m=+950.277163327" observedRunningTime="2025-12-10 14:40:02.18397684 +0000 UTC m=+951.753194490" watchObservedRunningTime="2025-12-10 14:40:02.186557242 +0000 UTC m=+951.755774872" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.198023 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" event={"ID":"2fa5c5a1-e3d6-424c-b52a-d01af15458cf","Type":"ContainerStarted","Data":"b6edd03b24228fd537385ab5d04a814486fe2d592dc565bfbf0695eba2fa9794"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.255236 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" event={"ID":"defc8bda-15a6-4432-8079-9606a8fdb2e9","Type":"ContainerStarted","Data":"eb33d000a8a71f33b2893048664c9a965a34825a9eeed41637bf28d70efc9f8f"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.255748 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.264024 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.303056 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" event={"ID":"79090a78-3da6-4b17-a875-fc8c2d534076","Type":"ContainerStarted","Data":"327c21b7b3120030eea2627511ff0a58fd25fe76cbb1bedb190a3ddd3cb47523"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.306048 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.323484 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.337968 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" event={"ID":"971c05b8-38ac-411a-bdc8-b40ed70daab5","Type":"ContainerStarted","Data":"a3fe2c1f23dc0bf3126d6d0538e74c36dc5478beaba86613556d80ce231c06f7"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.339232 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.347980 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.371260 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" event={"ID":"161aaad4-8b14-4a9c-b793-a7937e078aef","Type":"ContainerStarted","Data":"af1a677e82220a497673287460a8a90038bf2d628735d4b92fe89d8302515a27"} Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.372634 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.385209 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sbtv2" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.385355 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-pmnjz" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.392898 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.445842 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" podStartSLOduration=2.933816931 podStartE2EDuration="33.445815522s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.971105975 +0000 UTC m=+920.540323605" lastFinishedPulling="2025-12-10 14:40:01.483104566 +0000 UTC m=+951.052322196" observedRunningTime="2025-12-10 14:40:02.438143559 +0000 UTC m=+952.007361199" watchObservedRunningTime="2025-12-10 14:40:02.445815522 +0000 UTC m=+952.015033152" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.514530 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-t224h" podStartSLOduration=4.036291584 podStartE2EDuration="33.514513073s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.226913379 +0000 UTC m=+920.796131019" lastFinishedPulling="2025-12-10 14:40:00.705134878 +0000 UTC m=+950.274352508" observedRunningTime="2025-12-10 14:40:02.513927877 +0000 UTC m=+952.083145517" watchObservedRunningTime="2025-12-10 14:40:02.514513073 +0000 UTC m=+952.083730703" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.630821 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-z6f9c" podStartSLOduration=4.130250997 podStartE2EDuration="33.630801987s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.206572724 +0000 UTC m=+920.775790364" lastFinishedPulling="2025-12-10 14:40:00.707123724 +0000 UTC m=+950.276341354" observedRunningTime="2025-12-10 14:40:02.629035549 +0000 UTC m=+952.198253179" watchObservedRunningTime="2025-12-10 14:40:02.630801987 +0000 UTC m=+952.200019607" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.637901 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-t2wb8" podStartSLOduration=3.5845682500000002 podStartE2EDuration="33.637877604s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.497983605 +0000 UTC m=+920.067201235" lastFinishedPulling="2025-12-10 14:40:00.551292959 +0000 UTC m=+950.120510589" observedRunningTime="2025-12-10 14:40:02.593644704 +0000 UTC m=+952.162862354" watchObservedRunningTime="2025-12-10 14:40:02.637877604 +0000 UTC m=+952.207095234" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.690037 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-q8fkp" podStartSLOduration=3.406754234 podStartE2EDuration="33.690013004s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.422082494 +0000 UTC m=+919.991300124" lastFinishedPulling="2025-12-10 14:40:00.705341264 +0000 UTC m=+950.274558894" observedRunningTime="2025-12-10 14:40:02.665907394 +0000 UTC m=+952.235125024" watchObservedRunningTime="2025-12-10 14:40:02.690013004 +0000 UTC m=+952.259230634" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.828250 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-kds4v" podStartSLOduration=4.078677824 podStartE2EDuration="33.828227429s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.958392681 +0000 UTC m=+920.527610311" lastFinishedPulling="2025-12-10 14:40:00.707942286 +0000 UTC m=+950.277159916" observedRunningTime="2025-12-10 14:40:02.82720634 +0000 UTC m=+952.396424000" watchObservedRunningTime="2025-12-10 14:40:02.828227429 +0000 UTC m=+952.397445059" Dec 10 14:40:02 crc kubenswrapper[4847]: I1210 14:40:02.940923 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv"] Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.107369 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt"] Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.379253 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" event={"ID":"b1d011f5-5052-431d-9933-314f65930ce4","Type":"ContainerStarted","Data":"14bf797f0d8dc1a7847d5bafe49690c00b36c40d631d90f109af3f8d15451f5a"} Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.383727 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" event={"ID":"d78cab92-e99d-41fb-a35d-ff63d3ceba23","Type":"ContainerStarted","Data":"5f09726820a8f9b7bdf3d5b39b519f0e632d506937e4a0de79ae970c7411bb77"} Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.389841 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" event={"ID":"80cfb202-f181-4c1a-b9d6-a7cd036b93db","Type":"ContainerStarted","Data":"42ecf3ed1ff54c94d45dd01227b6e13546de8d68ce7d98e78dd39192c06a9dcb"} Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.393240 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:40:03 crc kubenswrapper[4847]: I1210 14:40:03.415178 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" podStartSLOduration=1.897430075 podStartE2EDuration="34.415154392s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.137095118 +0000 UTC m=+919.706312748" lastFinishedPulling="2025-12-10 14:40:02.654819435 +0000 UTC m=+952.224037065" observedRunningTime="2025-12-10 14:40:03.411030898 +0000 UTC m=+952.980248528" watchObservedRunningTime="2025-12-10 14:40:03.415154392 +0000 UTC m=+952.984372022" Dec 10 14:40:09 crc kubenswrapper[4847]: I1210 14:40:09.386681 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-rqjxj" Dec 10 14:40:09 crc kubenswrapper[4847]: I1210 14:40:09.824574 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gdhjq" Dec 10 14:40:15 crc kubenswrapper[4847]: I1210 14:40:15.476748 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" event={"ID":"3654ef6c-9eef-40ec-acd3-c101a817f345","Type":"ContainerStarted","Data":"1553621da49765093e225f1fd8fad4cdb0d3b0db3d84630d6399506959b3a5f3"} Dec 10 14:40:15 crc kubenswrapper[4847]: I1210 14:40:15.478946 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" event={"ID":"2fa5c5a1-e3d6-424c-b52a-d01af15458cf","Type":"ContainerStarted","Data":"11a57575ac6aceea3680335fe960b2aa4bc2ff88543c9fea875e25e1c9ba14a3"} Dec 10 14:40:15 crc kubenswrapper[4847]: I1210 14:40:15.480809 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-twm4t" Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.488039 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" event={"ID":"9f72b089-2518-4144-a558-bfb5c490d594","Type":"ContainerStarted","Data":"0e56b592ab21cf1bd53e904b5fc7ad6cf2246473df55f98f1e9a687a5baa6abf"} Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.489684 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" event={"ID":"9b0a2f2c-971b-40cc-8544-9842f7878d62","Type":"ContainerStarted","Data":"6028db6741a7ae7953786441b96b7fea451a22df8ef7a313f688ca4167a227b4"} Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.491011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" event={"ID":"b1d011f5-5052-431d-9933-314f65930ce4","Type":"ContainerStarted","Data":"ca8e7078c15cde4b640e64798b7a704adf1b5be6bc018970caaf5a30a8691b0d"} Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.492394 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" event={"ID":"df30724c-6ddb-4e68-8f57-31f8bb262678","Type":"ContainerStarted","Data":"c6a5bc98b2ca51e28ea94b2450e51979d5369b578d61d10e166cc8e871c1c129"} Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.493973 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" event={"ID":"c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90","Type":"ContainerStarted","Data":"bb28e0cbc75406ae2e2164ec2ddc7e0ec122081fa20239dd634a327f1f174c13"} Dec 10 14:40:16 crc kubenswrapper[4847]: I1210 14:40:16.497622 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" event={"ID":"0c5922f3-c544-4993-b978-4a035ec08e1b","Type":"ContainerStarted","Data":"4adaafd6dc9583c33ed58b6bf307f0ed262e05bbfeb4000406fec84a88a8b853"} Dec 10 14:40:17 crc kubenswrapper[4847]: E1210 14:40:17.791845 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage147552540/1\": happened during read: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48" Dec 10 14:40:17 crc kubenswrapper[4847]: E1210 14:40:17.792695 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f8cv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-84b575879f2swnv_openstack-operators(d78cab92-e99d-41fb-a35d-ff63d3ceba23): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage147552540/1\": happened during read: context canceled" logger="UnhandledError" Dec 10 14:40:17 crc kubenswrapper[4847]: E1210 14:40:17.929657 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage147552540/1\\\": happened during read: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" podUID="d78cab92-e99d-41fb-a35d-ff63d3ceba23" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.512301 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" event={"ID":"d78cab92-e99d-41fb-a35d-ff63d3ceba23","Type":"ContainerStarted","Data":"b70bf3b3b3487e34d6a20c7d75de2986145a094b250243e4996c76a8920cb659"} Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.512801 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.512874 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:40:18 crc kubenswrapper[4847]: E1210 14:40:18.513975 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" podUID="d78cab92-e99d-41fb-a35d-ff63d3ceba23" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.532465 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" podStartSLOduration=16.900749177 podStartE2EDuration="49.532445695s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.390028702 +0000 UTC m=+919.959246332" lastFinishedPulling="2025-12-10 14:40:03.02172522 +0000 UTC m=+952.590942850" observedRunningTime="2025-12-10 14:40:18.531469007 +0000 UTC m=+968.100686647" watchObservedRunningTime="2025-12-10 14:40:18.532445695 +0000 UTC m=+968.101663325" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.553346 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" podStartSLOduration=17.988806589 podStartE2EDuration="49.553325855s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.367502749 +0000 UTC m=+920.936720379" lastFinishedPulling="2025-12-10 14:40:02.932022015 +0000 UTC m=+952.501239645" observedRunningTime="2025-12-10 14:40:18.549095088 +0000 UTC m=+968.118312728" watchObservedRunningTime="2025-12-10 14:40:18.553325855 +0000 UTC m=+968.122543475" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.572103 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" podStartSLOduration=17.71771546 podStartE2EDuration="49.572086547s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.202201122 +0000 UTC m=+920.771418752" lastFinishedPulling="2025-12-10 14:40:03.056572199 +0000 UTC m=+952.625789839" observedRunningTime="2025-12-10 14:40:18.569604718 +0000 UTC m=+968.138822368" watchObservedRunningTime="2025-12-10 14:40:18.572086547 +0000 UTC m=+968.141304177" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.602980 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" podStartSLOduration=49.602955926 podStartE2EDuration="49.602955926s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:40:18.600174858 +0000 UTC m=+968.169392508" watchObservedRunningTime="2025-12-10 14:40:18.602955926 +0000 UTC m=+968.172173556" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.619983 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" podStartSLOduration=17.547143226 podStartE2EDuration="49.619962199s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.983062427 +0000 UTC m=+920.552280057" lastFinishedPulling="2025-12-10 14:40:03.0558814 +0000 UTC m=+952.625099030" observedRunningTime="2025-12-10 14:40:18.616594705 +0000 UTC m=+968.185812335" watchObservedRunningTime="2025-12-10 14:40:18.619962199 +0000 UTC m=+968.189179819" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.635048 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" podStartSLOduration=17.792001586 podStartE2EDuration="49.635031788s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:31.197984165 +0000 UTC m=+920.767201795" lastFinishedPulling="2025-12-10 14:40:03.041014367 +0000 UTC m=+952.610231997" observedRunningTime="2025-12-10 14:40:18.630727898 +0000 UTC m=+968.199945528" watchObservedRunningTime="2025-12-10 14:40:18.635031788 +0000 UTC m=+968.204249428" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.694934 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" podStartSLOduration=17.70728514 podStartE2EDuration="49.694912774s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.862777871 +0000 UTC m=+920.431995501" lastFinishedPulling="2025-12-10 14:40:02.850405505 +0000 UTC m=+952.419623135" observedRunningTime="2025-12-10 14:40:18.679839774 +0000 UTC m=+968.249057394" watchObservedRunningTime="2025-12-10 14:40:18.694912774 +0000 UTC m=+968.264130404" Dec 10 14:40:18 crc kubenswrapper[4847]: I1210 14:40:18.699208 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" podStartSLOduration=17.567328656 podStartE2EDuration="49.699196922s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:39:30.885571345 +0000 UTC m=+920.454788975" lastFinishedPulling="2025-12-10 14:40:03.017439611 +0000 UTC m=+952.586657241" observedRunningTime="2025-12-10 14:40:18.695938422 +0000 UTC m=+968.265156052" watchObservedRunningTime="2025-12-10 14:40:18.699196922 +0000 UTC m=+968.268414552" Dec 10 14:40:19 crc kubenswrapper[4847]: E1210 14:40:19.520431 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:9d539fb6b72f91cfc6200bb91b7c6dbaeab17c7711342dd3a9549c66762a2d48\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" podUID="d78cab92-e99d-41fb-a35d-ff63d3ceba23" Dec 10 14:40:19 crc kubenswrapper[4847]: I1210 14:40:19.693500 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:40:19 crc kubenswrapper[4847]: I1210 14:40:19.702109 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:40:19 crc kubenswrapper[4847]: I1210 14:40:19.794051 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:40:20 crc kubenswrapper[4847]: I1210 14:40:20.019283 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:40:20 crc kubenswrapper[4847]: I1210 14:40:20.021669 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-4dfs5" Dec 10 14:40:20 crc kubenswrapper[4847]: I1210 14:40:20.139039 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:40:20 crc kubenswrapper[4847]: I1210 14:40:20.141429 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8k92s" Dec 10 14:40:20 crc kubenswrapper[4847]: I1210 14:40:20.165783 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-4dhdf" Dec 10 14:40:21 crc kubenswrapper[4847]: I1210 14:40:21.983566 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:21 crc kubenswrapper[4847]: I1210 14:40:21.989922 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79b6c4fddb-8mzvt" Dec 10 14:40:29 crc kubenswrapper[4847]: I1210 14:40:29.686661 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-55zkq" Dec 10 14:40:29 crc kubenswrapper[4847]: I1210 14:40:29.695784 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-pdp9r" Dec 10 14:40:29 crc kubenswrapper[4847]: I1210 14:40:29.707729 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-sptmg" Dec 10 14:40:29 crc kubenswrapper[4847]: I1210 14:40:29.798438 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gjtlq" Dec 10 14:40:36 crc kubenswrapper[4847]: I1210 14:40:36.645122 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" event={"ID":"d78cab92-e99d-41fb-a35d-ff63d3ceba23","Type":"ContainerStarted","Data":"fe5ca9008122b2ef7a0b5b166d86d8bf90802a5b40218b0010877adc54720b29"} Dec 10 14:40:36 crc kubenswrapper[4847]: I1210 14:40:36.646128 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:40:36 crc kubenswrapper[4847]: I1210 14:40:36.674185 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" podStartSLOduration=34.275908836 podStartE2EDuration="1m7.674167409s" podCreationTimestamp="2025-12-10 14:39:29 +0000 UTC" firstStartedPulling="2025-12-10 14:40:03.021666009 +0000 UTC m=+952.590883639" lastFinishedPulling="2025-12-10 14:40:36.419924582 +0000 UTC m=+985.989142212" observedRunningTime="2025-12-10 14:40:36.669204622 +0000 UTC m=+986.238422272" watchObservedRunningTime="2025-12-10 14:40:36.674167409 +0000 UTC m=+986.243385039" Dec 10 14:40:41 crc kubenswrapper[4847]: I1210 14:40:41.724338 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f2swnv" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.485613 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.490192 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.495270 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.495699 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-tnk9x" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.496047 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.496299 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.505905 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.569493 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.572351 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.573682 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9tp2\" (UniqueName: \"kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.574121 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.575838 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.579512 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.675134 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.675233 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.675270 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5wmq\" (UniqueName: \"kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.675295 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.675337 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9tp2\" (UniqueName: \"kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.676247 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.698156 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9tp2\" (UniqueName: \"kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2\") pod \"dnsmasq-dns-675f4bcbfc-ldn5m\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.776111 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.776159 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5wmq\" (UniqueName: \"kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.776182 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.777074 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.777130 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.805647 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5wmq\" (UniqueName: \"kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq\") pod \"dnsmasq-dns-78dd6ddcc-xl5f5\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.847445 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:40:57 crc kubenswrapper[4847]: I1210 14:40:57.888369 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:40:58 crc kubenswrapper[4847]: I1210 14:40:58.253985 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:40:58 crc kubenswrapper[4847]: I1210 14:40:58.258229 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:40:58 crc kubenswrapper[4847]: I1210 14:40:58.350202 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:40:58 crc kubenswrapper[4847]: W1210 14:40:58.352790 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10069337_515f_4699_9cac_4577c3ec42a0.slice/crio-5ee829732e2c046a7c894bce71b9e98a0f62b4a67bb8f8ef6d1b1450f0a1bffd WatchSource:0}: Error finding container 5ee829732e2c046a7c894bce71b9e98a0f62b4a67bb8f8ef6d1b1450f0a1bffd: Status 404 returned error can't find the container with id 5ee829732e2c046a7c894bce71b9e98a0f62b4a67bb8f8ef6d1b1450f0a1bffd Dec 10 14:40:58 crc kubenswrapper[4847]: I1210 14:40:58.807204 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" event={"ID":"edd31212-9f11-4181-8e19-1494f548d759","Type":"ContainerStarted","Data":"843fa3c4dd4e5e6716e3eb8fcf58c502ea2b6abe91878450d6224753d886a2c3"} Dec 10 14:40:58 crc kubenswrapper[4847]: I1210 14:40:58.809210 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" event={"ID":"10069337-515f-4699-9cac-4577c3ec42a0","Type":"ContainerStarted","Data":"5ee829732e2c046a7c894bce71b9e98a0f62b4a67bb8f8ef6d1b1450f0a1bffd"} Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.468435 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.500501 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.502436 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.526139 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.627594 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.627755 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.627814 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ld6\" (UniqueName: \"kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.729148 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.729245 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.729277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ld6\" (UniqueName: \"kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.730372 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.731266 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.770946 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ld6\" (UniqueName: \"kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6\") pod \"dnsmasq-dns-666b6646f7-4f2tm\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.784744 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.798579 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.812168 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.840917 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.863121 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.962907 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscfb\" (UniqueName: \"kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.962970 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:00 crc kubenswrapper[4847]: I1210 14:41:00.963031 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.064631 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscfb\" (UniqueName: \"kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.064733 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.064816 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.065765 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.066757 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.089387 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscfb\" (UniqueName: \"kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb\") pod \"dnsmasq-dns-57d769cc4f-5pc46\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.165169 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.612296 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.615933 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.619070 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.619475 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-bbhz4" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.619619 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.620497 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.620622 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.620741 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.620866 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.631897 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671829 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvb2r\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671894 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671913 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671957 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671982 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.671998 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.672025 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.672045 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.672070 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.672089 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.672108 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774374 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvb2r\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774438 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774463 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774517 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774549 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774725 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774947 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.774992 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.775027 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.775048 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.775069 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.776428 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.776751 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.778317 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.779859 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.782468 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.788481 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.849153 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.849204 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.852229 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.905758 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.925579 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvb2r\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.928665 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.929851 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.935345 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.936633 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.941541 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " pod="openstack/rabbitmq-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.942250 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.942483 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.942657 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wvgrd" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.942778 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.944392 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.967668 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979263 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979311 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979363 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979399 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979423 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9vp2\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979514 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979574 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.979962 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.980037 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.980057 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:01 crc kubenswrapper[4847]: I1210 14:41:01.980125 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.082555 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.082662 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.082774 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.082811 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.083199 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.083246 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9vp2\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.083329 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.084812 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.084887 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.084954 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.084985 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.085008 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.085132 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.086015 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.086315 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.086463 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.086560 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.088252 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.090163 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.090564 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.090610 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.100021 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9vp2\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.112147 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.241652 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:41:02 crc kubenswrapper[4847]: I1210 14:41:02.293725 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.535570 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.538898 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.542343 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9kvbr" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.545221 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.548413 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.550482 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.550576 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.553481 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.611647 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.611969 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2gxz\" (UniqueName: \"kubernetes.io/projected/32d8cf5e-b11b-469d-8838-b27e201d95c0-kube-api-access-n2gxz\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.612073 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.612803 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.612900 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.612978 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.613083 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.613285 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715300 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715353 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715412 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715436 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2gxz\" (UniqueName: \"kubernetes.io/projected/32d8cf5e-b11b-469d-8838-b27e201d95c0-kube-api-access-n2gxz\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715467 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715482 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715497 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.715515 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.716224 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.716657 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.716690 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-kolla-config\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.716984 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-config-data-default\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.717988 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32d8cf5e-b11b-469d-8838-b27e201d95c0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.720202 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.731331 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32d8cf5e-b11b-469d-8838-b27e201d95c0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.742638 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.744886 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2gxz\" (UniqueName: \"kubernetes.io/projected/32d8cf5e-b11b-469d-8838-b27e201d95c0-kube-api-access-n2gxz\") pod \"openstack-galera-0\" (UID: \"32d8cf5e-b11b-469d-8838-b27e201d95c0\") " pod="openstack/openstack-galera-0" Dec 10 14:41:03 crc kubenswrapper[4847]: I1210 14:41:03.901909 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.898386 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.900171 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.902416 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8t9fs" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.902667 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.903542 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.903670 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.909536 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930776 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930896 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930918 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930942 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.930976 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4jdm\" (UniqueName: \"kubernetes.io/projected/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kube-api-access-z4jdm\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.931015 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:04 crc kubenswrapper[4847]: I1210 14:41:04.931036 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.031811 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.031878 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.031928 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4jdm\" (UniqueName: \"kubernetes.io/projected/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kube-api-access-z4jdm\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.031996 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032028 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032067 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032109 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032145 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032339 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.032792 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.033159 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.033529 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.033760 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4499b911-76a6-4ceb-8320-c87af3f1bd0a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.043642 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.044753 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4499b911-76a6-4ceb-8320-c87af3f1bd0a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.054545 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.055652 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4jdm\" (UniqueName: \"kubernetes.io/projected/4499b911-76a6-4ceb-8320-c87af3f1bd0a-kube-api-access-z4jdm\") pod \"openstack-cell1-galera-0\" (UID: \"4499b911-76a6-4ceb-8320-c87af3f1bd0a\") " pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.194178 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.195366 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.197285 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.197431 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nb99k" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.197791 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.230287 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.236246 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.335466 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.335816 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-config-data\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.335936 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58c5b\" (UniqueName: \"kubernetes.io/projected/fc672732-248d-4f66-88fd-3b98ce15e78d-kube-api-access-58c5b\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.335980 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.336023 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-kolla-config\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.437320 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58c5b\" (UniqueName: \"kubernetes.io/projected/fc672732-248d-4f66-88fd-3b98ce15e78d-kube-api-access-58c5b\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.437387 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.437422 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-kolla-config\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.437462 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.437568 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-config-data\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.438376 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-kolla-config\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.441866 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.450685 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc672732-248d-4f66-88fd-3b98ce15e78d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.464917 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58c5b\" (UniqueName: \"kubernetes.io/projected/fc672732-248d-4f66-88fd-3b98ce15e78d-kube-api-access-58c5b\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.776169 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc672732-248d-4f66-88fd-3b98ce15e78d-config-data\") pod \"memcached-0\" (UID: \"fc672732-248d-4f66-88fd-3b98ce15e78d\") " pod="openstack/memcached-0" Dec 10 14:41:05 crc kubenswrapper[4847]: I1210 14:41:05.810570 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.290282 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.300282 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.302950 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-dtkrx" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.313777 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.468882 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnmg5\" (UniqueName: \"kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5\") pod \"kube-state-metrics-0\" (UID: \"4d586df7-072d-4e8f-b2c2-250c31c20a29\") " pod="openstack/kube-state-metrics-0" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.570870 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnmg5\" (UniqueName: \"kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5\") pod \"kube-state-metrics-0\" (UID: \"4d586df7-072d-4e8f-b2c2-250c31c20a29\") " pod="openstack/kube-state-metrics-0" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.589729 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnmg5\" (UniqueName: \"kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5\") pod \"kube-state-metrics-0\" (UID: \"4d586df7-072d-4e8f-b2c2-250c31c20a29\") " pod="openstack/kube-state-metrics-0" Dec 10 14:41:07 crc kubenswrapper[4847]: I1210 14:41:07.622920 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.338890 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.944207 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.949029 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.952760 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.952952 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.953034 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.953069 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.953101 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-xcwdm" Dec 10 14:41:10 crc kubenswrapper[4847]: I1210 14:41:10.961981 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131617 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131690 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131751 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-config\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131788 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g4cg\" (UniqueName: \"kubernetes.io/projected/26654ec0-c7fa-44da-9be9-8d5218addac5-kube-api-access-9g4cg\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131834 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131907 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.131978 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234215 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234258 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234302 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234347 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234382 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234402 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234423 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-config\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234447 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g4cg\" (UniqueName: \"kubernetes.io/projected/26654ec0-c7fa-44da-9be9-8d5218addac5-kube-api-access-9g4cg\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.234660 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.236136 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.237207 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-config\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.241319 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26654ec0-c7fa-44da-9be9-8d5218addac5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.242579 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.242581 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.245020 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26654ec0-c7fa-44da-9be9-8d5218addac5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.251992 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g4cg\" (UniqueName: \"kubernetes.io/projected/26654ec0-c7fa-44da-9be9-8d5218addac5-kube-api-access-9g4cg\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.267952 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"26654ec0-c7fa-44da-9be9-8d5218addac5\") " pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.277840 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.475740 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2xjh7"] Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.476944 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.480248 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.480795 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dq5b6" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.481002 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.493136 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7"] Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.556513 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rnkkd"] Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.559009 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.569245 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rnkkd"] Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.644748 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-combined-ca-bundle\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.644817 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b962e42f-e21b-4016-9587-111fb6eb68ac-scripts\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.644839 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.644864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.645003 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-log-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.645055 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-ovn-controller-tls-certs\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.645121 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95cwk\" (UniqueName: \"kubernetes.io/projected/b962e42f-e21b-4016-9587-111fb6eb68ac-kube-api-access-95cwk\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746069 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746139 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-etc-ovs\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746169 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746199 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-log-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746223 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0ee780-f899-462e-89dd-e353a5790288-scripts\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746327 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-lib\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746429 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-run\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746476 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-ovn-controller-tls-certs\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746562 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clwrk\" (UniqueName: \"kubernetes.io/projected/6f0ee780-f899-462e-89dd-e353a5790288-kube-api-access-clwrk\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746598 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95cwk\" (UniqueName: \"kubernetes.io/projected/b962e42f-e21b-4016-9587-111fb6eb68ac-kube-api-access-95cwk\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746644 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-log\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746683 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746705 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-combined-ca-bundle\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746770 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-run\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746796 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b962e42f-e21b-4016-9587-111fb6eb68ac-scripts\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.746976 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b962e42f-e21b-4016-9587-111fb6eb68ac-var-log-ovn\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.748942 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b962e42f-e21b-4016-9587-111fb6eb68ac-scripts\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.752569 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-combined-ca-bundle\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.753079 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/b962e42f-e21b-4016-9587-111fb6eb68ac-ovn-controller-tls-certs\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.770297 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95cwk\" (UniqueName: \"kubernetes.io/projected/b962e42f-e21b-4016-9587-111fb6eb68ac-kube-api-access-95cwk\") pod \"ovn-controller-2xjh7\" (UID: \"b962e42f-e21b-4016-9587-111fb6eb68ac\") " pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.801796 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848697 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clwrk\" (UniqueName: \"kubernetes.io/projected/6f0ee780-f899-462e-89dd-e353a5790288-kube-api-access-clwrk\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848785 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-log\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848852 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-etc-ovs\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848875 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0ee780-f899-462e-89dd-e353a5790288-scripts\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848893 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-lib\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.848978 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-run\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.849108 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-run\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.849202 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-log\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.849313 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-var-lib\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.849370 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6f0ee780-f899-462e-89dd-e353a5790288-etc-ovs\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.853747 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f0ee780-f899-462e-89dd-e353a5790288-scripts\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.865387 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clwrk\" (UniqueName: \"kubernetes.io/projected/6f0ee780-f899-462e-89dd-e353a5790288-kube-api-access-clwrk\") pod \"ovn-controller-ovs-rnkkd\" (UID: \"6f0ee780-f899-462e-89dd-e353a5790288\") " pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:11 crc kubenswrapper[4847]: I1210 14:41:11.883647 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.724521 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.725784 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.731374 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.731987 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.732299 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-cbs6x" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.734783 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.750086 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910332 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910391 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910497 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910530 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbqxq\" (UniqueName: \"kubernetes.io/projected/beaec882-8047-4348-bcf5-a45eed4dba20-kube-api-access-kbqxq\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910576 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-config\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910626 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910651 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:14 crc kubenswrapper[4847]: I1210 14:41:14.910678 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012451 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012500 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbqxq\" (UniqueName: \"kubernetes.io/projected/beaec882-8047-4348-bcf5-a45eed4dba20-kube-api-access-kbqxq\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012534 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-config\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012559 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012578 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012595 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012767 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.012801 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.014995 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.015528 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-config\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.015943 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.017526 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.019734 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beaec882-8047-4348-bcf5-a45eed4dba20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.020796 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.027510 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/beaec882-8047-4348-bcf5-a45eed4dba20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.031193 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbqxq\" (UniqueName: \"kubernetes.io/projected/beaec882-8047-4348-bcf5-a45eed4dba20-kube-api-access-kbqxq\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.040171 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"beaec882-8047-4348-bcf5-a45eed4dba20\") " pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:15 crc kubenswrapper[4847]: I1210 14:41:15.062834 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:17 crc kubenswrapper[4847]: W1210 14:41:17.336629 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1a52184_c799_484f_bb34_9c4f42e15e83.slice/crio-adf066884edd4383e89fd3acc42a3b8531fab15dd0cdf2dd2669dcfaf30b0917 WatchSource:0}: Error finding container adf066884edd4383e89fd3acc42a3b8531fab15dd0cdf2dd2669dcfaf30b0917: Status 404 returned error can't find the container with id adf066884edd4383e89fd3acc42a3b8531fab15dd0cdf2dd2669dcfaf30b0917 Dec 10 14:41:17 crc kubenswrapper[4847]: I1210 14:41:17.788405 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:41:17 crc kubenswrapper[4847]: I1210 14:41:17.848093 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:18 crc kubenswrapper[4847]: I1210 14:41:18.053311 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" event={"ID":"f1a52184-c799-484f-bb34-9c4f42e15e83","Type":"ContainerStarted","Data":"adf066884edd4383e89fd3acc42a3b8531fab15dd0cdf2dd2669dcfaf30b0917"} Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.248860 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.249335 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z9tp2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-ldn5m_openstack(edd31212-9f11-4181-8e19-1494f548d759): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.250873 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" podUID="edd31212-9f11-4181-8e19-1494f548d759" Dec 10 14:41:18 crc kubenswrapper[4847]: W1210 14:41:18.256386 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d33f956_6feb_4c03_88f9_95185d25f781.slice/crio-8bcf30f8b59d24440c4be23ee88e7d90cb412343081859d7fb1fbd6b4f4a58d2 WatchSource:0}: Error finding container 8bcf30f8b59d24440c4be23ee88e7d90cb412343081859d7fb1fbd6b4f4a58d2: Status 404 returned error can't find the container with id 8bcf30f8b59d24440c4be23ee88e7d90cb412343081859d7fb1fbd6b4f4a58d2 Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.273840 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.274004 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t5wmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-xl5f5_openstack(10069337-515f-4699-9cac-4577c3ec42a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:41:18 crc kubenswrapper[4847]: E1210 14:41:18.276423 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" podUID="10069337-515f-4699-9cac-4577c3ec42a0" Dec 10 14:41:18 crc kubenswrapper[4847]: I1210 14:41:18.825023 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:41:18 crc kubenswrapper[4847]: W1210 14:41:18.832545 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d586df7_072d_4e8f_b2c2_250c31c20a29.slice/crio-f5e244c99713928a7b216d10294ace5a596215d47fa35ad025f6306dd78a8367 WatchSource:0}: Error finding container f5e244c99713928a7b216d10294ace5a596215d47fa35ad025f6306dd78a8367: Status 404 returned error can't find the container with id f5e244c99713928a7b216d10294ace5a596215d47fa35ad025f6306dd78a8367 Dec 10 14:41:18 crc kubenswrapper[4847]: I1210 14:41:18.851680 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 14:41:18 crc kubenswrapper[4847]: W1210 14:41:18.857988 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f5ccab8_d00d_43e4_938c_e49a1cdfc3d9.slice/crio-7e012cfc4d4b7c04524b283e49bb5fa18e47a46ea2b1517feb4553e55dac581b WatchSource:0}: Error finding container 7e012cfc4d4b7c04524b283e49bb5fa18e47a46ea2b1517feb4553e55dac581b: Status 404 returned error can't find the container with id 7e012cfc4d4b7c04524b283e49bb5fa18e47a46ea2b1517feb4553e55dac581b Dec 10 14:41:18 crc kubenswrapper[4847]: I1210 14:41:18.860939 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:41:18 crc kubenswrapper[4847]: W1210 14:41:18.866485 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32d8cf5e_b11b_469d_8838_b27e201d95c0.slice/crio-5e0a01a01b9c5c8f06d0f99dfd67c86a7f1542cb13d42949f21c94cd01e1ce41 WatchSource:0}: Error finding container 5e0a01a01b9c5c8f06d0f99dfd67c86a7f1542cb13d42949f21c94cd01e1ce41: Status 404 returned error can't find the container with id 5e0a01a01b9c5c8f06d0f99dfd67c86a7f1542cb13d42949f21c94cd01e1ce41 Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.045451 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7"] Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.051540 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.063792 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7" event={"ID":"b962e42f-e21b-4016-9587-111fb6eb68ac","Type":"ContainerStarted","Data":"9157c46c269ad76babe387bf1d8f9630436e573dc4485a28b16e720ef77698ed"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.065120 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerStarted","Data":"8bcf30f8b59d24440c4be23ee88e7d90cb412343081859d7fb1fbd6b4f4a58d2"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.067123 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" event={"ID":"c626935d-b06a-4245-9e2b-0f2010f00bc0","Type":"ContainerStarted","Data":"a6366cae2bf871258743908627898de0443ef5da321e7a71cbac5dd41fa43812"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.070252 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d586df7-072d-4e8f-b2c2-250c31c20a29","Type":"ContainerStarted","Data":"f5e244c99713928a7b216d10294ace5a596215d47fa35ad025f6306dd78a8367"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.075417 4847 generic.go:334] "Generic (PLEG): container finished" podID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerID="c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d" exitCode=0 Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.075688 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" event={"ID":"f1a52184-c799-484f-bb34-9c4f42e15e83","Type":"ContainerDied","Data":"c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.080754 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"32d8cf5e-b11b-469d-8838-b27e201d95c0","Type":"ContainerStarted","Data":"5e0a01a01b9c5c8f06d0f99dfd67c86a7f1542cb13d42949f21c94cd01e1ce41"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.086653 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerStarted","Data":"7e012cfc4d4b7c04524b283e49bb5fa18e47a46ea2b1517feb4553e55dac581b"} Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.156594 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.200289 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.353498 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rnkkd"] Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.461698 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.609390 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc\") pod \"10069337-515f-4699-9cac-4577c3ec42a0\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.609773 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5wmq\" (UniqueName: \"kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq\") pod \"10069337-515f-4699-9cac-4577c3ec42a0\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.609818 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config\") pod \"10069337-515f-4699-9cac-4577c3ec42a0\" (UID: \"10069337-515f-4699-9cac-4577c3ec42a0\") " Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.610065 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10069337-515f-4699-9cac-4577c3ec42a0" (UID: "10069337-515f-4699-9cac-4577c3ec42a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.610345 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config" (OuterVolumeSpecName: "config") pod "10069337-515f-4699-9cac-4577c3ec42a0" (UID: "10069337-515f-4699-9cac-4577c3ec42a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.610701 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.610749 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10069337-515f-4699-9cac-4577c3ec42a0-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.614361 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq" (OuterVolumeSpecName: "kube-api-access-t5wmq") pod "10069337-515f-4699-9cac-4577c3ec42a0" (UID: "10069337-515f-4699-9cac-4577c3ec42a0"). InnerVolumeSpecName "kube-api-access-t5wmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.659506 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.712341 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5wmq\" (UniqueName: \"kubernetes.io/projected/10069337-515f-4699-9cac-4577c3ec42a0-kube-api-access-t5wmq\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.813309 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config\") pod \"edd31212-9f11-4181-8e19-1494f548d759\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.813449 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9tp2\" (UniqueName: \"kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2\") pod \"edd31212-9f11-4181-8e19-1494f548d759\" (UID: \"edd31212-9f11-4181-8e19-1494f548d759\") " Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.813861 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config" (OuterVolumeSpecName: "config") pod "edd31212-9f11-4181-8e19-1494f548d759" (UID: "edd31212-9f11-4181-8e19-1494f548d759"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.816609 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2" (OuterVolumeSpecName: "kube-api-access-z9tp2") pod "edd31212-9f11-4181-8e19-1494f548d759" (UID: "edd31212-9f11-4181-8e19-1494f548d759"). InnerVolumeSpecName "kube-api-access-z9tp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.915319 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edd31212-9f11-4181-8e19-1494f548d759-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:19 crc kubenswrapper[4847]: I1210 14:41:19.915354 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9tp2\" (UniqueName: \"kubernetes.io/projected/edd31212-9f11-4181-8e19-1494f548d759-kube-api-access-z9tp2\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.050040 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.096625 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rnkkd" event={"ID":"6f0ee780-f899-462e-89dd-e353a5790288","Type":"ContainerStarted","Data":"71031449f014b388dbcb64407c7427169a516663a508cd7d1d9e65b4885d0879"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.098202 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4499b911-76a6-4ceb-8320-c87af3f1bd0a","Type":"ContainerStarted","Data":"55fd611247e5651dfe81b8f9f3d3238f1f8a2d4653ca0814a6bad391c47a4029"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.099368 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" event={"ID":"edd31212-9f11-4181-8e19-1494f548d759","Type":"ContainerDied","Data":"843fa3c4dd4e5e6716e3eb8fcf58c502ea2b6abe91878450d6224753d886a2c3"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.099433 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-ldn5m" Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.100479 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"beaec882-8047-4348-bcf5-a45eed4dba20","Type":"ContainerStarted","Data":"535cb471452b5d923834437b80da9d829e85fcd016be0ab84c6b93c4cd674d5d"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.103629 4847 generic.go:334] "Generic (PLEG): container finished" podID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerID="5cabca541076c892b3958799bd20191127fce0fc669d98986f7e091b71b70b59" exitCode=0 Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.103739 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" event={"ID":"c626935d-b06a-4245-9e2b-0f2010f00bc0","Type":"ContainerDied","Data":"5cabca541076c892b3958799bd20191127fce0fc669d98986f7e091b71b70b59"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.105304 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fc672732-248d-4f66-88fd-3b98ce15e78d","Type":"ContainerStarted","Data":"f81fb0b76bdb379603b8bec8eaf8cf035f43f875ab4dceb9d7f276aee62fd69c"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.106665 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" event={"ID":"10069337-515f-4699-9cac-4577c3ec42a0","Type":"ContainerDied","Data":"5ee829732e2c046a7c894bce71b9e98a0f62b4a67bb8f8ef6d1b1450f0a1bffd"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.106707 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-xl5f5" Dec 10 14:41:20 crc kubenswrapper[4847]: W1210 14:41:20.109785 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26654ec0_c7fa_44da_9be9_8d5218addac5.slice/crio-3b3d485a30ba8a74574ec7e4742ac8c398f5724b0fc91db738ca6f5d74b3293d WatchSource:0}: Error finding container 3b3d485a30ba8a74574ec7e4742ac8c398f5724b0fc91db738ca6f5d74b3293d: Status 404 returned error can't find the container with id 3b3d485a30ba8a74574ec7e4742ac8c398f5724b0fc91db738ca6f5d74b3293d Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.118052 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" event={"ID":"f1a52184-c799-484f-bb34-9c4f42e15e83","Type":"ContainerStarted","Data":"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e"} Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.118357 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.150684 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" podStartSLOduration=19.130268826 podStartE2EDuration="20.150654748s" podCreationTimestamp="2025-12-10 14:41:00 +0000 UTC" firstStartedPulling="2025-12-10 14:41:17.343068211 +0000 UTC m=+1026.912285841" lastFinishedPulling="2025-12-10 14:41:18.363454133 +0000 UTC m=+1027.932671763" observedRunningTime="2025-12-10 14:41:20.139399707 +0000 UTC m=+1029.708617337" watchObservedRunningTime="2025-12-10 14:41:20.150654748 +0000 UTC m=+1029.719872398" Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.179738 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.187125 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-ldn5m"] Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.207532 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.219535 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-xl5f5"] Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.769303 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10069337-515f-4699-9cac-4577c3ec42a0" path="/var/lib/kubelet/pods/10069337-515f-4699-9cac-4577c3ec42a0/volumes" Dec 10 14:41:20 crc kubenswrapper[4847]: I1210 14:41:20.770948 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd31212-9f11-4181-8e19-1494f548d759" path="/var/lib/kubelet/pods/edd31212-9f11-4181-8e19-1494f548d759/volumes" Dec 10 14:41:21 crc kubenswrapper[4847]: I1210 14:41:21.125692 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"26654ec0-c7fa-44da-9be9-8d5218addac5","Type":"ContainerStarted","Data":"3b3d485a30ba8a74574ec7e4742ac8c398f5724b0fc91db738ca6f5d74b3293d"} Dec 10 14:41:26 crc kubenswrapper[4847]: I1210 14:41:26.166699 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:26 crc kubenswrapper[4847]: I1210 14:41:26.228287 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:28 crc kubenswrapper[4847]: I1210 14:41:28.190794 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" event={"ID":"c626935d-b06a-4245-9e2b-0f2010f00bc0","Type":"ContainerStarted","Data":"00143c0dae6df68b21422196c533659d71e6ee8f5a37e6ef096448d74f03b88e"} Dec 10 14:41:28 crc kubenswrapper[4847]: I1210 14:41:28.191315 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:28 crc kubenswrapper[4847]: I1210 14:41:28.190912 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="dnsmasq-dns" containerID="cri-o://00143c0dae6df68b21422196c533659d71e6ee8f5a37e6ef096448d74f03b88e" gracePeriod=10 Dec 10 14:41:28 crc kubenswrapper[4847]: I1210 14:41:28.195442 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"32d8cf5e-b11b-469d-8838-b27e201d95c0","Type":"ContainerStarted","Data":"67036a82643398b7664d758a419782d94dd8d0675de715f32637f0251694acef"} Dec 10 14:41:28 crc kubenswrapper[4847]: I1210 14:41:28.237900 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" podStartSLOduration=27.56397742 podStartE2EDuration="28.237877618s" podCreationTimestamp="2025-12-10 14:41:00 +0000 UTC" firstStartedPulling="2025-12-10 14:41:18.261188341 +0000 UTC m=+1027.830405971" lastFinishedPulling="2025-12-10 14:41:18.935088539 +0000 UTC m=+1028.504306169" observedRunningTime="2025-12-10 14:41:28.209583517 +0000 UTC m=+1037.778801147" watchObservedRunningTime="2025-12-10 14:41:28.237877618 +0000 UTC m=+1037.807095248" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.204042 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fc672732-248d-4f66-88fd-3b98ce15e78d","Type":"ContainerStarted","Data":"bfc2e063b82d06fd29b00250c9345571365a0c084276d5afec96ef8ae7c295fb"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.204589 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.206476 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d586df7-072d-4e8f-b2c2-250c31c20a29","Type":"ContainerStarted","Data":"45fbd8f086bedf06905f3c444934fb69237df6a3ad9d5d1758c7eea701f28abb"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.206586 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.209194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"26654ec0-c7fa-44da-9be9-8d5218addac5","Type":"ContainerStarted","Data":"261ee89e7e9a236290740c692067b2d065faf7c100120954a15e00cbafb9a171"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.210756 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7" event={"ID":"b962e42f-e21b-4016-9587-111fb6eb68ac","Type":"ContainerStarted","Data":"1aaa52ae5f7387e50b16b11b29b44864f02d716ab043b9e9944256b6e668bcb8"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.210849 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-2xjh7" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.212153 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"beaec882-8047-4348-bcf5-a45eed4dba20","Type":"ContainerStarted","Data":"1806c8c982025bb1d6c9e52d3c82a3c416971bdea1aaf457ad572e05164723c3"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.214108 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerStarted","Data":"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.215877 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerStarted","Data":"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.224892 4847 generic.go:334] "Generic (PLEG): container finished" podID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerID="00143c0dae6df68b21422196c533659d71e6ee8f5a37e6ef096448d74f03b88e" exitCode=0 Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.224971 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" event={"ID":"c626935d-b06a-4245-9e2b-0f2010f00bc0","Type":"ContainerDied","Data":"00143c0dae6df68b21422196c533659d71e6ee8f5a37e6ef096448d74f03b88e"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.225730 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" event={"ID":"c626935d-b06a-4245-9e2b-0f2010f00bc0","Type":"ContainerDied","Data":"a6366cae2bf871258743908627898de0443ef5da321e7a71cbac5dd41fa43812"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.225751 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6366cae2bf871258743908627898de0443ef5da321e7a71cbac5dd41fa43812" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.228490 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=16.10940893 podStartE2EDuration="24.228467718s" podCreationTimestamp="2025-12-10 14:41:05 +0000 UTC" firstStartedPulling="2025-12-10 14:41:19.250462134 +0000 UTC m=+1028.819679764" lastFinishedPulling="2025-12-10 14:41:27.369520932 +0000 UTC m=+1036.938738552" observedRunningTime="2025-12-10 14:41:29.223233773 +0000 UTC m=+1038.792451423" watchObservedRunningTime="2025-12-10 14:41:29.228467718 +0000 UTC m=+1038.797685348" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.229012 4847 generic.go:334] "Generic (PLEG): container finished" podID="6f0ee780-f899-462e-89dd-e353a5790288" containerID="de7f8f83cdb42280ca0d2bf6b6bc21eed5151b36f4bba80fe4ce1ee8f9c378ea" exitCode=0 Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.229219 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rnkkd" event={"ID":"6f0ee780-f899-462e-89dd-e353a5790288","Type":"ContainerDied","Data":"de7f8f83cdb42280ca0d2bf6b6bc21eed5151b36f4bba80fe4ce1ee8f9c378ea"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.231938 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.231985 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4499b911-76a6-4ceb-8320-c87af3f1bd0a","Type":"ContainerStarted","Data":"9dafdb0cb223a4f045cc977a106dc8b9a1ca225a8c795c8ee0a1f42601e6512a"} Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.242588 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2xjh7" podStartSLOduration=9.954882724 podStartE2EDuration="18.242571667s" podCreationTimestamp="2025-12-10 14:41:11 +0000 UTC" firstStartedPulling="2025-12-10 14:41:19.054917067 +0000 UTC m=+1028.624134697" lastFinishedPulling="2025-12-10 14:41:27.34260601 +0000 UTC m=+1036.911823640" observedRunningTime="2025-12-10 14:41:29.240962952 +0000 UTC m=+1038.810180592" watchObservedRunningTime="2025-12-10 14:41:29.242571667 +0000 UTC m=+1038.811789297" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.299832 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.791883025 podStartE2EDuration="22.299796696s" podCreationTimestamp="2025-12-10 14:41:07 +0000 UTC" firstStartedPulling="2025-12-10 14:41:18.834948316 +0000 UTC m=+1028.404165956" lastFinishedPulling="2025-12-10 14:41:27.342861997 +0000 UTC m=+1036.912079627" observedRunningTime="2025-12-10 14:41:29.28075885 +0000 UTC m=+1038.849976480" watchObservedRunningTime="2025-12-10 14:41:29.299796696 +0000 UTC m=+1038.869014336" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.366553 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config\") pod \"c626935d-b06a-4245-9e2b-0f2010f00bc0\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.366645 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc\") pod \"c626935d-b06a-4245-9e2b-0f2010f00bc0\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.366819 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ld6\" (UniqueName: \"kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6\") pod \"c626935d-b06a-4245-9e2b-0f2010f00bc0\" (UID: \"c626935d-b06a-4245-9e2b-0f2010f00bc0\") " Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.378693 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6" (OuterVolumeSpecName: "kube-api-access-j9ld6") pod "c626935d-b06a-4245-9e2b-0f2010f00bc0" (UID: "c626935d-b06a-4245-9e2b-0f2010f00bc0"). InnerVolumeSpecName "kube-api-access-j9ld6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.409596 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config" (OuterVolumeSpecName: "config") pod "c626935d-b06a-4245-9e2b-0f2010f00bc0" (UID: "c626935d-b06a-4245-9e2b-0f2010f00bc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.432615 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c626935d-b06a-4245-9e2b-0f2010f00bc0" (UID: "c626935d-b06a-4245-9e2b-0f2010f00bc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.468886 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.468922 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c626935d-b06a-4245-9e2b-0f2010f00bc0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:29 crc kubenswrapper[4847]: I1210 14:41:29.468934 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ld6\" (UniqueName: \"kubernetes.io/projected/c626935d-b06a-4245-9e2b-0f2010f00bc0-kube-api-access-j9ld6\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.258796 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-4f2tm" Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.259676 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rnkkd" event={"ID":"6f0ee780-f899-462e-89dd-e353a5790288","Type":"ContainerStarted","Data":"de64e5ef294062c13cc3c0ae83b8c95f1049e8c88d338825344943970e505611"} Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.259732 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rnkkd" event={"ID":"6f0ee780-f899-462e-89dd-e353a5790288","Type":"ContainerStarted","Data":"6a1e3cbb87009ec3ea0a61b0a31f536936973516e35cadd4beba13b3ade19e97"} Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.259997 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.260051 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.286381 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rnkkd" podStartSLOduration=11.326327425 podStartE2EDuration="19.286362454s" podCreationTimestamp="2025-12-10 14:41:11 +0000 UTC" firstStartedPulling="2025-12-10 14:41:19.362472435 +0000 UTC m=+1028.931690065" lastFinishedPulling="2025-12-10 14:41:27.322507464 +0000 UTC m=+1036.891725094" observedRunningTime="2025-12-10 14:41:30.282773295 +0000 UTC m=+1039.851990925" watchObservedRunningTime="2025-12-10 14:41:30.286362454 +0000 UTC m=+1039.855580084" Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.304046 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.311618 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-4f2tm"] Dec 10 14:41:30 crc kubenswrapper[4847]: I1210 14:41:30.777687 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" path="/var/lib/kubelet/pods/c626935d-b06a-4245-9e2b-0f2010f00bc0/volumes" Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.275157 4847 generic.go:334] "Generic (PLEG): container finished" podID="32d8cf5e-b11b-469d-8838-b27e201d95c0" containerID="67036a82643398b7664d758a419782d94dd8d0675de715f32637f0251694acef" exitCode=0 Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.275232 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"32d8cf5e-b11b-469d-8838-b27e201d95c0","Type":"ContainerDied","Data":"67036a82643398b7664d758a419782d94dd8d0675de715f32637f0251694acef"} Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.277047 4847 generic.go:334] "Generic (PLEG): container finished" podID="4499b911-76a6-4ceb-8320-c87af3f1bd0a" containerID="9dafdb0cb223a4f045cc977a106dc8b9a1ca225a8c795c8ee0a1f42601e6512a" exitCode=0 Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.277084 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4499b911-76a6-4ceb-8320-c87af3f1bd0a","Type":"ContainerDied","Data":"9dafdb0cb223a4f045cc977a106dc8b9a1ca225a8c795c8ee0a1f42601e6512a"} Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.278899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"beaec882-8047-4348-bcf5-a45eed4dba20","Type":"ContainerStarted","Data":"cab334427b645b2e7437dfa6240601a28e61f1c80235cb925cd6b33316690c1d"} Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.281372 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"26654ec0-c7fa-44da-9be9-8d5218addac5","Type":"ContainerStarted","Data":"75b13452de5ef28a15d94735530b57015dc1d3158cc6a98103fec6e5c99deb55"} Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.320282 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.473457748 podStartE2EDuration="23.320260537s" podCreationTimestamp="2025-12-10 14:41:09 +0000 UTC" firstStartedPulling="2025-12-10 14:41:20.11228746 +0000 UTC m=+1029.681505090" lastFinishedPulling="2025-12-10 14:41:31.959090239 +0000 UTC m=+1041.528307879" observedRunningTime="2025-12-10 14:41:32.311807894 +0000 UTC m=+1041.881025534" watchObservedRunningTime="2025-12-10 14:41:32.320260537 +0000 UTC m=+1041.889478167" Dec 10 14:41:32 crc kubenswrapper[4847]: I1210 14:41:32.357201 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.615772756 podStartE2EDuration="19.357182997s" podCreationTimestamp="2025-12-10 14:41:13 +0000 UTC" firstStartedPulling="2025-12-10 14:41:19.208338551 +0000 UTC m=+1028.777556181" lastFinishedPulling="2025-12-10 14:41:31.949748792 +0000 UTC m=+1041.518966422" observedRunningTime="2025-12-10 14:41:32.351979093 +0000 UTC m=+1041.921196743" watchObservedRunningTime="2025-12-10 14:41:32.357182997 +0000 UTC m=+1041.926400627" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.063828 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.112766 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.290580 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"32d8cf5e-b11b-469d-8838-b27e201d95c0","Type":"ContainerStarted","Data":"58afc57ed07cedf6336377e223eea0b4f7bec2fa9a75a5c723028eb0b1f7d5ce"} Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.294375 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4499b911-76a6-4ceb-8320-c87af3f1bd0a","Type":"ContainerStarted","Data":"cd05799174108ebf6d6e73242fa80f24b58a0b0b13bf400780fcc5410fff289c"} Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.294638 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.316106 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.842664333 podStartE2EDuration="31.316081872s" podCreationTimestamp="2025-12-10 14:41:02 +0000 UTC" firstStartedPulling="2025-12-10 14:41:18.869193041 +0000 UTC m=+1028.438410671" lastFinishedPulling="2025-12-10 14:41:27.34261058 +0000 UTC m=+1036.911828210" observedRunningTime="2025-12-10 14:41:33.311017692 +0000 UTC m=+1042.880235362" watchObservedRunningTime="2025-12-10 14:41:33.316081872 +0000 UTC m=+1042.885299522" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.339382 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.340810 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.029682745 podStartE2EDuration="30.340787704s" podCreationTimestamp="2025-12-10 14:41:03 +0000 UTC" firstStartedPulling="2025-12-10 14:41:19.06085106 +0000 UTC m=+1028.630068680" lastFinishedPulling="2025-12-10 14:41:27.371956009 +0000 UTC m=+1036.941173639" observedRunningTime="2025-12-10 14:41:33.331292201 +0000 UTC m=+1042.900509881" watchObservedRunningTime="2025-12-10 14:41:33.340787704 +0000 UTC m=+1042.910005354" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.592140 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:33 crc kubenswrapper[4847]: E1210 14:41:33.592829 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="dnsmasq-dns" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.592859 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="dnsmasq-dns" Dec 10 14:41:33 crc kubenswrapper[4847]: E1210 14:41:33.592893 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="init" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.592905 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="init" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.593175 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c626935d-b06a-4245-9e2b-0f2010f00bc0" containerName="dnsmasq-dns" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.594690 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.597849 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.605767 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.643103 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-n4rbc"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.644249 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.646417 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658422 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658479 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovn-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658532 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx55b\" (UniqueName: \"kubernetes.io/projected/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-kube-api-access-dx55b\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658595 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658616 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-config\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658645 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-combined-ca-bundle\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658667 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658728 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgskx\" (UniqueName: \"kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658789 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovs-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.658859 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.669912 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n4rbc"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760115 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760170 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-config\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760198 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-combined-ca-bundle\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760222 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760258 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgskx\" (UniqueName: \"kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760278 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovs-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760317 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760384 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760406 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovn-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.760455 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx55b\" (UniqueName: \"kubernetes.io/projected/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-kube-api-access-dx55b\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.761926 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.762222 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovs-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.762283 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-ovn-rundir\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.762316 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-config\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.762825 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.763021 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.767918 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-combined-ca-bundle\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.779106 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx55b\" (UniqueName: \"kubernetes.io/projected/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-kube-api-access-dx55b\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.779664 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f080a9d-d07e-4a5e-b782-245d4c7b31cf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-n4rbc\" (UID: \"4f080a9d-d07e-4a5e-b782-245d4c7b31cf\") " pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.790766 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgskx\" (UniqueName: \"kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx\") pod \"dnsmasq-dns-6bc7876d45-6cwf8\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.795186 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.796086 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.810187 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.812033 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.814243 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.828933 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.861964 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.862300 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.862324 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc76d\" (UniqueName: \"kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.862417 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.862454 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.903576 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.903805 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.961648 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-n4rbc" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.970082 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.970168 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.970189 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.970205 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc76d\" (UniqueName: \"kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.970273 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.971449 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.971483 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.973570 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.976115 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:33 crc kubenswrapper[4847]: I1210 14:41:33.990371 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc76d\" (UniqueName: \"kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d\") pod \"dnsmasq-dns-8554648995-gnqxb\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:34 crc kubenswrapper[4847]: I1210 14:41:34.192692 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:34 crc kubenswrapper[4847]: I1210 14:41:34.276386 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:34 crc kubenswrapper[4847]: W1210 14:41:34.278799 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604698e0_4162_4edf_aaba_a663ad51b4c4.slice/crio-effc580968d65f9ef798f15b706fb43224224721f8d3b602cb1db580ee56867e WatchSource:0}: Error finding container effc580968d65f9ef798f15b706fb43224224721f8d3b602cb1db580ee56867e: Status 404 returned error can't find the container with id effc580968d65f9ef798f15b706fb43224224721f8d3b602cb1db580ee56867e Dec 10 14:41:34 crc kubenswrapper[4847]: I1210 14:41:34.304105 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" event={"ID":"604698e0-4162-4edf-aaba-a663ad51b4c4","Type":"ContainerStarted","Data":"effc580968d65f9ef798f15b706fb43224224721f8d3b602cb1db580ee56867e"} Dec 10 14:41:34 crc kubenswrapper[4847]: I1210 14:41:34.411058 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-n4rbc"] Dec 10 14:41:34 crc kubenswrapper[4847]: W1210 14:41:34.418319 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f080a9d_d07e_4a5e_b782_245d4c7b31cf.slice/crio-64e425f99b0dbe6c082c6cbba53eb17e2460d047e27443e0080096a997b962b5 WatchSource:0}: Error finding container 64e425f99b0dbe6c082c6cbba53eb17e2460d047e27443e0080096a997b962b5: Status 404 returned error can't find the container with id 64e425f99b0dbe6c082c6cbba53eb17e2460d047e27443e0080096a997b962b5 Dec 10 14:41:34 crc kubenswrapper[4847]: E1210 14:41:34.633415 4847 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604698e0_4162_4edf_aaba_a663ad51b4c4.slice/crio-conmon-a05d01948cfc4a915ed13b32592d29f103ed69f6105508537de62d94fb846a81.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604698e0_4162_4edf_aaba_a663ad51b4c4.slice/crio-a05d01948cfc4a915ed13b32592d29f103ed69f6105508537de62d94fb846a81.scope\": RecentStats: unable to find data in memory cache]" Dec 10 14:41:34 crc kubenswrapper[4847]: I1210 14:41:34.679636 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:34 crc kubenswrapper[4847]: W1210 14:41:34.682375 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod924719e0_0b9e_4156_9d4f_bb1f8940a0e5.slice/crio-7e9a6330dde7fb99488bc0a5c2d0cb3269b53f846e1c4cf6d09e1361bf1c24bd WatchSource:0}: Error finding container 7e9a6330dde7fb99488bc0a5c2d0cb3269b53f846e1c4cf6d09e1361bf1c24bd: Status 404 returned error can't find the container with id 7e9a6330dde7fb99488bc0a5c2d0cb3269b53f846e1c4cf6d09e1361bf1c24bd Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.231350 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.231775 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.278824 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.316809 4847 generic.go:334] "Generic (PLEG): container finished" podID="604698e0-4162-4edf-aaba-a663ad51b4c4" containerID="a05d01948cfc4a915ed13b32592d29f103ed69f6105508537de62d94fb846a81" exitCode=0 Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.316931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" event={"ID":"604698e0-4162-4edf-aaba-a663ad51b4c4","Type":"ContainerDied","Data":"a05d01948cfc4a915ed13b32592d29f103ed69f6105508537de62d94fb846a81"} Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.317975 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.319833 4847 generic.go:334] "Generic (PLEG): container finished" podID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerID="0b83924f033ccdc66e04c426a3da0c4880a36a5a11b4562e7b5862538e4bdbd6" exitCode=0 Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.319977 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gnqxb" event={"ID":"924719e0-0b9e-4156-9d4f-bb1f8940a0e5","Type":"ContainerDied","Data":"0b83924f033ccdc66e04c426a3da0c4880a36a5a11b4562e7b5862538e4bdbd6"} Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.320020 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gnqxb" event={"ID":"924719e0-0b9e-4156-9d4f-bb1f8940a0e5","Type":"ContainerStarted","Data":"7e9a6330dde7fb99488bc0a5c2d0cb3269b53f846e1c4cf6d09e1361bf1c24bd"} Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.324068 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n4rbc" event={"ID":"4f080a9d-d07e-4a5e-b782-245d4c7b31cf","Type":"ContainerStarted","Data":"cb2b4b9b2b30ecf44426420bc598bc5f746bdeb688a71535461d4a50e452f890"} Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.324118 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-n4rbc" event={"ID":"4f080a9d-d07e-4a5e-b782-245d4c7b31cf","Type":"ContainerStarted","Data":"64e425f99b0dbe6c082c6cbba53eb17e2460d047e27443e0080096a997b962b5"} Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.325368 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.371939 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-n4rbc" podStartSLOduration=2.371910601 podStartE2EDuration="2.371910601s" podCreationTimestamp="2025-12-10 14:41:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:41:35.370754038 +0000 UTC m=+1044.939971668" watchObservedRunningTime="2025-12-10 14:41:35.371910601 +0000 UTC m=+1044.941128231" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.382528 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.662061 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.665310 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.670345 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.670410 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.670601 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nkrz6" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.671212 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.678020 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.789839 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799195 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799309 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-config\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799338 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799363 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799403 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-scripts\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799428 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbv6z\" (UniqueName: \"kubernetes.io/projected/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-kube-api-access-fbv6z\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.799652 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.812648 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.900961 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config\") pod \"604698e0-4162-4edf-aaba-a663ad51b4c4\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901069 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgskx\" (UniqueName: \"kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx\") pod \"604698e0-4162-4edf-aaba-a663ad51b4c4\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901148 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb\") pod \"604698e0-4162-4edf-aaba-a663ad51b4c4\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901250 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc\") pod \"604698e0-4162-4edf-aaba-a663ad51b4c4\" (UID: \"604698e0-4162-4edf-aaba-a663ad51b4c4\") " Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901518 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901576 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-config\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901605 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901627 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901678 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-scripts\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901699 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbv6z\" (UniqueName: \"kubernetes.io/projected/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-kube-api-access-fbv6z\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.901784 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.903919 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-scripts\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.904087 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.904546 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-config\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.910832 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.913728 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.913934 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx" (OuterVolumeSpecName: "kube-api-access-cgskx") pod "604698e0-4162-4edf-aaba-a663ad51b4c4" (UID: "604698e0-4162-4edf-aaba-a663ad51b4c4"). InnerVolumeSpecName "kube-api-access-cgskx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.914842 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.923778 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "604698e0-4162-4edf-aaba-a663ad51b4c4" (UID: "604698e0-4162-4edf-aaba-a663ad51b4c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.927732 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbv6z\" (UniqueName: \"kubernetes.io/projected/a4d0db06-c1bd-4af6-b78e-a1703d914c6d-kube-api-access-fbv6z\") pod \"ovn-northd-0\" (UID: \"a4d0db06-c1bd-4af6-b78e-a1703d914c6d\") " pod="openstack/ovn-northd-0" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.928561 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config" (OuterVolumeSpecName: "config") pod "604698e0-4162-4edf-aaba-a663ad51b4c4" (UID: "604698e0-4162-4edf-aaba-a663ad51b4c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:35 crc kubenswrapper[4847]: I1210 14:41:35.932729 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "604698e0-4162-4edf-aaba-a663ad51b4c4" (UID: "604698e0-4162-4edf-aaba-a663ad51b4c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.002448 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.003522 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.003582 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.003596 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgskx\" (UniqueName: \"kubernetes.io/projected/604698e0-4162-4edf-aaba-a663ad51b4c4-kube-api-access-cgskx\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.003609 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/604698e0-4162-4edf-aaba-a663ad51b4c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.332985 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.333257 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-6cwf8" event={"ID":"604698e0-4162-4edf-aaba-a663ad51b4c4","Type":"ContainerDied","Data":"effc580968d65f9ef798f15b706fb43224224721f8d3b602cb1db580ee56867e"} Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.333440 4847 scope.go:117] "RemoveContainer" containerID="a05d01948cfc4a915ed13b32592d29f103ed69f6105508537de62d94fb846a81" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.335374 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gnqxb" event={"ID":"924719e0-0b9e-4156-9d4f-bb1f8940a0e5","Type":"ContainerStarted","Data":"b7fe827045d1c4a9f7280321332204ea627c2f4180b8dba87336f8698b765023"} Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.335609 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.377297 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-gnqxb" podStartSLOduration=3.377274798 podStartE2EDuration="3.377274798s" podCreationTimestamp="2025-12-10 14:41:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:41:36.370359697 +0000 UTC m=+1045.939577337" watchObservedRunningTime="2025-12-10 14:41:36.377274798 +0000 UTC m=+1045.946492428" Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.431786 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.438634 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-6cwf8"] Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.486025 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 14:41:36 crc kubenswrapper[4847]: I1210 14:41:36.772140 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604698e0-4162-4edf-aaba-a663ad51b4c4" path="/var/lib/kubelet/pods/604698e0-4162-4edf-aaba-a663ad51b4c4/volumes" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.344828 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a4d0db06-c1bd-4af6-b78e-a1703d914c6d","Type":"ContainerStarted","Data":"3fd2840f9d6fc7e969d59b63c6dcc8d93d698ed6efd48bc144a93e34a0295c62"} Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.627304 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.702470 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.746617 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:41:37 crc kubenswrapper[4847]: E1210 14:41:37.747000 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604698e0-4162-4edf-aaba-a663ad51b4c4" containerName="init" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.747018 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="604698e0-4162-4edf-aaba-a663ad51b4c4" containerName="init" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.747161 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="604698e0-4162-4edf-aaba-a663ad51b4c4" containerName="init" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.747965 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.770102 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.836030 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.836088 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.836111 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.836142 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.836264 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2dvg\" (UniqueName: \"kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.937864 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2dvg\" (UniqueName: \"kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.937941 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.937968 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.937989 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.938015 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.938975 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.939656 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.939747 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.940304 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:37 crc kubenswrapper[4847]: I1210 14:41:37.955425 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2dvg\" (UniqueName: \"kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg\") pod \"dnsmasq-dns-b8fbc5445-57vhg\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.071974 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.352986 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-gnqxb" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="dnsmasq-dns" containerID="cri-o://b7fe827045d1c4a9f7280321332204ea627c2f4180b8dba87336f8698b765023" gracePeriod=10 Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.522516 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:41:38 crc kubenswrapper[4847]: W1210 14:41:38.531889 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfa565dd_ab76_4e5b_a464_369a60ccd1d3.slice/crio-6f584d4636d43a5d94dcc25d9d6ca3f3a1387d1eddf11b57321f0177d9348ded WatchSource:0}: Error finding container 6f584d4636d43a5d94dcc25d9d6ca3f3a1387d1eddf11b57321f0177d9348ded: Status 404 returned error can't find the container with id 6f584d4636d43a5d94dcc25d9d6ca3f3a1387d1eddf11b57321f0177d9348ded Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.814737 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.820061 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.821907 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.821910 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.823212 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8qqf9" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.823389 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.840245 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.953813 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.953886 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f2bd\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-kube-api-access-4f2bd\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.953992 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-lock\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.954020 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:38 crc kubenswrapper[4847]: I1210 14:41:38.954037 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-cache\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.055893 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-lock\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.055973 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.055991 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-cache\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.056028 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.056068 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f2bd\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-kube-api-access-4f2bd\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.056728 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.061007 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.061048 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.061107 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:41:39.561086459 +0000 UTC m=+1049.130304089 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.139274 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-lock\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.140409 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/88c96cdf-8002-4829-9f8c-3abfc8315722-cache\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.163365 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f2bd\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-kube-api-access-4f2bd\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.175001 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.363146 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" event={"ID":"cfa565dd-ab76-4e5b-a464-369a60ccd1d3","Type":"ContainerStarted","Data":"6f584d4636d43a5d94dcc25d9d6ca3f3a1387d1eddf11b57321f0177d9348ded"} Dec 10 14:41:39 crc kubenswrapper[4847]: I1210 14:41:39.564171 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.564338 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.564353 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:39 crc kubenswrapper[4847]: E1210 14:41:39.564405 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:41:40.564386489 +0000 UTC m=+1050.133604119 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:40 crc kubenswrapper[4847]: I1210 14:41:40.582902 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:40 crc kubenswrapper[4847]: E1210 14:41:40.583064 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:40 crc kubenswrapper[4847]: E1210 14:41:40.583329 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:40 crc kubenswrapper[4847]: E1210 14:41:40.583386 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:41:42.583370442 +0000 UTC m=+1052.152588072 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.616102 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:42 crc kubenswrapper[4847]: E1210 14:41:42.616331 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:42 crc kubenswrapper[4847]: E1210 14:41:42.617450 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:42 crc kubenswrapper[4847]: E1210 14:41:42.617562 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:41:46.617527413 +0000 UTC m=+1056.186745083 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.785679 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hkv2r"] Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.786775 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.789112 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.789217 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.789386 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.798614 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hkv2r"] Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922551 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922682 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922756 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922801 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922829 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxxrx\" (UniqueName: \"kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.922980 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:42 crc kubenswrapper[4847]: I1210 14:41:42.923074 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024765 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024824 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024863 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024889 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxxrx\" (UniqueName: \"kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024916 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024934 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.024984 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.025918 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.025931 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.026343 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.030811 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.032418 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.034011 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.042797 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxxrx\" (UniqueName: \"kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx\") pod \"swift-ring-rebalance-hkv2r\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.109022 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:41:43 crc kubenswrapper[4847]: I1210 14:41:43.580062 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hkv2r"] Dec 10 14:41:44 crc kubenswrapper[4847]: I1210 14:41:44.195492 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-gnqxb" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 10 14:41:45 crc kubenswrapper[4847]: I1210 14:41:45.098976 4847 generic.go:334] "Generic (PLEG): container finished" podID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerID="b7fe827045d1c4a9f7280321332204ea627c2f4180b8dba87336f8698b765023" exitCode=0 Dec 10 14:41:45 crc kubenswrapper[4847]: I1210 14:41:45.099053 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gnqxb" event={"ID":"924719e0-0b9e-4156-9d4f-bb1f8940a0e5","Type":"ContainerDied","Data":"b7fe827045d1c4a9f7280321332204ea627c2f4180b8dba87336f8698b765023"} Dec 10 14:41:46 crc kubenswrapper[4847]: I1210 14:41:46.114559 4847 generic.go:334] "Generic (PLEG): container finished" podID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerID="495c457f7386b14df806c517ee4578323d1bf0eee51d60f631890d0a1c759fbc" exitCode=0 Dec 10 14:41:46 crc kubenswrapper[4847]: I1210 14:41:46.114634 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" event={"ID":"cfa565dd-ab76-4e5b-a464-369a60ccd1d3","Type":"ContainerDied","Data":"495c457f7386b14df806c517ee4578323d1bf0eee51d60f631890d0a1c759fbc"} Dec 10 14:41:46 crc kubenswrapper[4847]: I1210 14:41:46.117482 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hkv2r" event={"ID":"daaeae82-74d4-4d18-8178-58ca7246b605","Type":"ContainerStarted","Data":"213dd121d779bcd20af6b7c1c43cdaa2dc8743d4ad6b706bb0443ae7400d9939"} Dec 10 14:41:46 crc kubenswrapper[4847]: I1210 14:41:46.691336 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:46 crc kubenswrapper[4847]: E1210 14:41:46.691568 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:46 crc kubenswrapper[4847]: E1210 14:41:46.691595 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:46 crc kubenswrapper[4847]: E1210 14:41:46.691659 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:41:54.691637096 +0000 UTC m=+1064.260854746 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:47 crc kubenswrapper[4847]: E1210 14:41:47.345820 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage551153743/1\": happened during read: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified" Dec 10 14:41:47 crc kubenswrapper[4847]: E1210 14:41:47.350135 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-northd,Image:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,Command:[/usr/bin/ovn-northd],Args:[-vfile:off -vconsole:info --n-threads=1 --ovnnb-db=ssl:ovsdbserver-nb-0.openstack.svc.cluster.local:6641 --ovnsb-db=ssl:ovsdbserver-sb-0.openstack.svc.cluster.local:6642 --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d7h4h58fh657h697h5f7h567h645h665h57bh5bch548h54dhddh55ch679h664h559h7h586h78h8bh64ch64dh5b7h668h66ch645hb7h65bh658hbq,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:certs,Value:n4h574hdch578h677h694h6dh95h558h685h5dbh7ch7dh56bh657h55h5bbh8hf6h648hbfh54bh5d5h57bh54bh5cfh689h66hbchbfh7dh5bdq,ValueFrom:nil,},EnvVar{Name:certs_metrics,Value:n66fh665h7fh5cdh566hf9h6h68h98h656h5dh548h8dh585h5c6h86h558h678h686h547h546h679h588hf8h644h85h9bh9bh594h688h5f5hdbq,ValueFrom:nil,},EnvVar{Name:ovnnorthd-config,Value:n5c8h7ch56bh8dh8hc4h5dch9dh68h6bhb7h598h549h5dbh66fh6bh5b4h5cch5d6h55ch57fhfch588h89h5ddh5d6h65bh65bh8dhc4h67dh569q,ValueFrom:nil,},EnvVar{Name:ovnnorthd-scripts,Value:n664hd8h66ch58dh64hc9h66bhd4h558h697h67bh557hdch664h567h669h555h696h556h556h5fh5bh569hbh665h9dh4h9bh564hc8h5b7h5c4q,ValueFrom:nil,},EnvVar{Name:tls-ca-bundle.pem,Value:n5c4hd4h54dh6fh58bh565hb7hbfh566h657h5f6h575h684h5dbh579h5dch5cdh5c6h5d6h688h69h6fh54bh5c7h87h7dh69h9fh64dh57ch59bh56fq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fbv6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-northd-0_openstack(a4d0db06-c1bd-4af6-b78e-a1703d914c6d): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage551153743/1\": happened during read: context canceled" logger="UnhandledError" Dec 10 14:41:47 crc kubenswrapper[4847]: E1210 14:41:47.584482 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage551153743/1\\\": happened during read: context canceled\"" pod="openstack/ovn-northd-0" podUID="a4d0db06-c1bd-4af6-b78e-a1703d914c6d" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.732113 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.811725 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config\") pod \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.811859 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc\") pod \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.811926 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc76d\" (UniqueName: \"kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d\") pod \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.811994 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb\") pod \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.812030 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb\") pod \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\" (UID: \"924719e0-0b9e-4156-9d4f-bb1f8940a0e5\") " Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.817806 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d" (OuterVolumeSpecName: "kube-api-access-gc76d") pod "924719e0-0b9e-4156-9d4f-bb1f8940a0e5" (UID: "924719e0-0b9e-4156-9d4f-bb1f8940a0e5"). InnerVolumeSpecName "kube-api-access-gc76d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.846593 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.855547 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config" (OuterVolumeSpecName: "config") pod "924719e0-0b9e-4156-9d4f-bb1f8940a0e5" (UID: "924719e0-0b9e-4156-9d4f-bb1f8940a0e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.860035 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "924719e0-0b9e-4156-9d4f-bb1f8940a0e5" (UID: "924719e0-0b9e-4156-9d4f-bb1f8940a0e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.873264 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "924719e0-0b9e-4156-9d4f-bb1f8940a0e5" (UID: "924719e0-0b9e-4156-9d4f-bb1f8940a0e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.884070 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "924719e0-0b9e-4156-9d4f-bb1f8940a0e5" (UID: "924719e0-0b9e-4156-9d4f-bb1f8940a0e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.914125 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc76d\" (UniqueName: \"kubernetes.io/projected/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-kube-api-access-gc76d\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.914361 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.914373 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.914382 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.914391 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924719e0-0b9e-4156-9d4f-bb1f8940a0e5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:47 crc kubenswrapper[4847]: I1210 14:41:47.934741 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.146316 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" event={"ID":"cfa565dd-ab76-4e5b-a464-369a60ccd1d3","Type":"ContainerStarted","Data":"9c52e790d767fb1767e08540f6c5f108fd6b5f0f456449f96dcc25484c438a42"} Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.146447 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.149309 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-gnqxb" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.149315 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-gnqxb" event={"ID":"924719e0-0b9e-4156-9d4f-bb1f8940a0e5","Type":"ContainerDied","Data":"7e9a6330dde7fb99488bc0a5c2d0cb3269b53f846e1c4cf6d09e1361bf1c24bd"} Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.149395 4847 scope.go:117] "RemoveContainer" containerID="b7fe827045d1c4a9f7280321332204ea627c2f4180b8dba87336f8698b765023" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.150878 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a4d0db06-c1bd-4af6-b78e-a1703d914c6d","Type":"ContainerStarted","Data":"3e06963657264955defc770fdaf401c495888a203b0c588698927502d22e036d"} Dec 10 14:41:48 crc kubenswrapper[4847]: E1210 14:41:48.152474 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="a4d0db06-c1bd-4af6-b78e-a1703d914c6d" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.169419 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" podStartSLOduration=11.16939842 podStartE2EDuration="11.16939842s" podCreationTimestamp="2025-12-10 14:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:41:48.163472196 +0000 UTC m=+1057.732689837" watchObservedRunningTime="2025-12-10 14:41:48.16939842 +0000 UTC m=+1057.738616050" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.171374 4847 scope.go:117] "RemoveContainer" containerID="0b83924f033ccdc66e04c426a3da0c4880a36a5a11b4562e7b5862538e4bdbd6" Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.233934 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.247173 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-gnqxb"] Dec 10 14:41:48 crc kubenswrapper[4847]: I1210 14:41:48.785768 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" path="/var/lib/kubelet/pods/924719e0-0b9e-4156-9d4f-bb1f8940a0e5/volumes" Dec 10 14:41:49 crc kubenswrapper[4847]: E1210 14:41:49.162884 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="a4d0db06-c1bd-4af6-b78e-a1703d914c6d" Dec 10 14:41:50 crc kubenswrapper[4847]: I1210 14:41:50.048961 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 10 14:41:50 crc kubenswrapper[4847]: I1210 14:41:50.135305 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 10 14:41:52 crc kubenswrapper[4847]: I1210 14:41:52.189632 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hkv2r" event={"ID":"daaeae82-74d4-4d18-8178-58ca7246b605","Type":"ContainerStarted","Data":"b014b0d3566ba7ce4bf058ed5832d3c4f04c9dac5047c94ebf7cdc0492c05c90"} Dec 10 14:41:52 crc kubenswrapper[4847]: I1210 14:41:52.211103 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hkv2r" podStartSLOduration=2.574530035 podStartE2EDuration="10.211083977s" podCreationTimestamp="2025-12-10 14:41:42 +0000 UTC" firstStartedPulling="2025-12-10 14:41:43.595260718 +0000 UTC m=+1053.164478348" lastFinishedPulling="2025-12-10 14:41:51.23181466 +0000 UTC m=+1060.801032290" observedRunningTime="2025-12-10 14:41:52.206687586 +0000 UTC m=+1061.775905236" watchObservedRunningTime="2025-12-10 14:41:52.211083977 +0000 UTC m=+1061.780301607" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.074128 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.144212 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.144555 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="dnsmasq-dns" containerID="cri-o://75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e" gracePeriod=10 Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.617121 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.717336 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc\") pod \"f1a52184-c799-484f-bb34-9c4f42e15e83\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.717386 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config\") pod \"f1a52184-c799-484f-bb34-9c4f42e15e83\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.717454 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pscfb\" (UniqueName: \"kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb\") pod \"f1a52184-c799-484f-bb34-9c4f42e15e83\" (UID: \"f1a52184-c799-484f-bb34-9c4f42e15e83\") " Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.725409 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb" (OuterVolumeSpecName: "kube-api-access-pscfb") pod "f1a52184-c799-484f-bb34-9c4f42e15e83" (UID: "f1a52184-c799-484f-bb34-9c4f42e15e83"). InnerVolumeSpecName "kube-api-access-pscfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.756255 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config" (OuterVolumeSpecName: "config") pod "f1a52184-c799-484f-bb34-9c4f42e15e83" (UID: "f1a52184-c799-484f-bb34-9c4f42e15e83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.759120 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1a52184-c799-484f-bb34-9c4f42e15e83" (UID: "f1a52184-c799-484f-bb34-9c4f42e15e83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.819617 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pscfb\" (UniqueName: \"kubernetes.io/projected/f1a52184-c799-484f-bb34-9c4f42e15e83-kube-api-access-pscfb\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.819659 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:53 crc kubenswrapper[4847]: I1210 14:41:53.819670 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1a52184-c799-484f-bb34-9c4f42e15e83-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.206635 4847 generic.go:334] "Generic (PLEG): container finished" podID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerID="75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e" exitCode=0 Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.206692 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" event={"ID":"f1a52184-c799-484f-bb34-9c4f42e15e83","Type":"ContainerDied","Data":"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e"} Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.207041 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" event={"ID":"f1a52184-c799-484f-bb34-9c4f42e15e83","Type":"ContainerDied","Data":"adf066884edd4383e89fd3acc42a3b8531fab15dd0cdf2dd2669dcfaf30b0917"} Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.207064 4847 scope.go:117] "RemoveContainer" containerID="75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.206743 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5pc46" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.238222 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.244421 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5pc46"] Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.245140 4847 scope.go:117] "RemoveContainer" containerID="c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.261139 4847 scope.go:117] "RemoveContainer" containerID="75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e" Dec 10 14:41:54 crc kubenswrapper[4847]: E1210 14:41:54.261499 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e\": container with ID starting with 75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e not found: ID does not exist" containerID="75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.261548 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e"} err="failed to get container status \"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e\": rpc error: code = NotFound desc = could not find container \"75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e\": container with ID starting with 75ad01640622e7a470274d38eeaef6e64f2a9d7a65b52e457e3ab9a025cc797e not found: ID does not exist" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.261572 4847 scope.go:117] "RemoveContainer" containerID="c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d" Dec 10 14:41:54 crc kubenswrapper[4847]: E1210 14:41:54.261949 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d\": container with ID starting with c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d not found: ID does not exist" containerID="c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.261979 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d"} err="failed to get container status \"c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d\": rpc error: code = NotFound desc = could not find container \"c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d\": container with ID starting with c973ac9573f1662a1317f32fafcf795b4ec9071859969a21275840f3cfb18b2d not found: ID does not exist" Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.737823 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:41:54 crc kubenswrapper[4847]: E1210 14:41:54.738075 4847 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 14:41:54 crc kubenswrapper[4847]: E1210 14:41:54.738117 4847 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 14:41:54 crc kubenswrapper[4847]: E1210 14:41:54.738198 4847 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift podName:88c96cdf-8002-4829-9f8c-3abfc8315722 nodeName:}" failed. No retries permitted until 2025-12-10 14:42:10.738169283 +0000 UTC m=+1080.307386923 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift") pod "swift-storage-0" (UID: "88c96cdf-8002-4829-9f8c-3abfc8315722") : configmap "swift-ring-files" not found Dec 10 14:41:54 crc kubenswrapper[4847]: I1210 14:41:54.769957 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" path="/var/lib/kubelet/pods/f1a52184-c799-484f-bb34-9c4f42e15e83/volumes" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.168767 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2595-account-create-update-hxnfg"] Dec 10 14:41:55 crc kubenswrapper[4847]: E1210 14:41:55.169380 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="init" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169396 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="init" Dec 10 14:41:55 crc kubenswrapper[4847]: E1210 14:41:55.169422 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169428 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: E1210 14:41:55.169441 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="init" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169447 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="init" Dec 10 14:41:55 crc kubenswrapper[4847]: E1210 14:41:55.169463 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169470 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169624 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="924719e0-0b9e-4156-9d4f-bb1f8940a0e5" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.169645 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1a52184-c799-484f-bb34-9c4f42e15e83" containerName="dnsmasq-dns" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.170194 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.173536 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.183975 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2595-account-create-update-hxnfg"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.249350 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.249404 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.253027 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-sj7rt"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.254231 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.261411 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sj7rt"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.351618 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74bqb\" (UniqueName: \"kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.351665 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.351731 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.351750 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.352370 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.370976 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65\") pod \"keystone-2595-account-create-update-hxnfg\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.413365 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2l5wx"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.414580 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.420631 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2l5wx"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.453017 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k59m\" (UniqueName: \"kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.453178 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74bqb\" (UniqueName: \"kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.453201 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.453241 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.453803 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.470145 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74bqb\" (UniqueName: \"kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb\") pod \"keystone-db-create-sj7rt\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.519398 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a9e3-account-create-update-8s2rf"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.520824 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.523112 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.524212 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.536653 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a9e3-account-create-update-8s2rf"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.554929 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q2lt\" (UniqueName: \"kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.555123 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.555227 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.555315 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k59m\" (UniqueName: \"kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.556145 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.576187 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k59m\" (UniqueName: \"kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m\") pod \"placement-db-create-2l5wx\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.577657 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.665502 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q2lt\" (UniqueName: \"kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.665955 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.666960 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.696323 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q2lt\" (UniqueName: \"kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt\") pod \"placement-a9e3-account-create-update-8s2rf\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.741011 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.946145 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jgdv2"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.947316 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.953666 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jgdv2"] Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.964787 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:55 crc kubenswrapper[4847]: I1210 14:41:55.990150 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2595-account-create-update-hxnfg"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.072540 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-260c-account-create-update-cnfqw"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.073833 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.076308 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.080879 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-260c-account-create-update-cnfqw"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.080887 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2v9t\" (UniqueName: \"kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.080984 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.159200 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-sj7rt"] Dec 10 14:41:56 crc kubenswrapper[4847]: W1210 14:41:56.161241 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06fc695e_403f_4858_adfc_0ed1f8d414c7.slice/crio-094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41 WatchSource:0}: Error finding container 094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41: Status 404 returned error can't find the container with id 094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41 Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.182911 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2v9t\" (UniqueName: \"kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.182972 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.183012 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz475\" (UniqueName: \"kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.183057 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.184354 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.206873 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2v9t\" (UniqueName: \"kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t\") pod \"glance-db-create-jgdv2\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.244466 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2595-account-create-update-hxnfg" event={"ID":"995e809e-0a8d-4b1e-acd8-b64571571651","Type":"ContainerStarted","Data":"dab5b9227e9761db9e739272db5d831ff23121f2dafb0cd2ff29f353e3e20b4c"} Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.244518 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2595-account-create-update-hxnfg" event={"ID":"995e809e-0a8d-4b1e-acd8-b64571571651","Type":"ContainerStarted","Data":"0768579e60569b8539bf8c1be9edbb431a4233f7d9501466c3948d25d72c71fa"} Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.246834 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sj7rt" event={"ID":"06fc695e-403f-4858-adfc-0ed1f8d414c7","Type":"ContainerStarted","Data":"094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41"} Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.271174 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.284861 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz475\" (UniqueName: \"kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.284938 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.285676 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.303465 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz475\" (UniqueName: \"kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475\") pod \"glance-260c-account-create-update-cnfqw\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.407350 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.441116 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-2595-account-create-update-hxnfg" podStartSLOduration=1.441093372 podStartE2EDuration="1.441093372s" podCreationTimestamp="2025-12-10 14:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:41:56.261920016 +0000 UTC m=+1065.831137666" watchObservedRunningTime="2025-12-10 14:41:56.441093372 +0000 UTC m=+1066.010311002" Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.446959 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2l5wx"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.511405 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a9e3-account-create-update-8s2rf"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.782333 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jgdv2"] Dec 10 14:41:56 crc kubenswrapper[4847]: I1210 14:41:56.936419 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-260c-account-create-update-cnfqw"] Dec 10 14:41:56 crc kubenswrapper[4847]: W1210 14:41:56.948235 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17e3cce4_5609_460e_b5b1_8e8a17962bc9.slice/crio-5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654 WatchSource:0}: Error finding container 5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654: Status 404 returned error can't find the container with id 5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.255799 4847 generic.go:334] "Generic (PLEG): container finished" podID="995e809e-0a8d-4b1e-acd8-b64571571651" containerID="dab5b9227e9761db9e739272db5d831ff23121f2dafb0cd2ff29f353e3e20b4c" exitCode=0 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.255889 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2595-account-create-update-hxnfg" event={"ID":"995e809e-0a8d-4b1e-acd8-b64571571651","Type":"ContainerDied","Data":"dab5b9227e9761db9e739272db5d831ff23121f2dafb0cd2ff29f353e3e20b4c"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.267053 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-260c-account-create-update-cnfqw" event={"ID":"17e3cce4-5609-460e-b5b1-8e8a17962bc9","Type":"ContainerStarted","Data":"5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.272357 4847 generic.go:334] "Generic (PLEG): container finished" podID="8711478e-be68-44d2-9471-225c0008b966" containerID="b06a4da59be924173fb8a1d5b7e8c78a26a0174a4fd641e2a61605c216ee093e" exitCode=0 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.272436 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2l5wx" event={"ID":"8711478e-be68-44d2-9471-225c0008b966","Type":"ContainerDied","Data":"b06a4da59be924173fb8a1d5b7e8c78a26a0174a4fd641e2a61605c216ee093e"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.272465 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2l5wx" event={"ID":"8711478e-be68-44d2-9471-225c0008b966","Type":"ContainerStarted","Data":"c73b68c4955b64b204968dcc29c52465151912ad9c66d27dcb74940aad1b07e3"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.274208 4847 generic.go:334] "Generic (PLEG): container finished" podID="abac02be-a75e-4ba2-ac00-4cb338681fec" containerID="365bb82dbca333a1c0be98734a242477ddd0290bb75bb639cfbc566d1dd8b5d5" exitCode=0 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.274512 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jgdv2" event={"ID":"abac02be-a75e-4ba2-ac00-4cb338681fec","Type":"ContainerDied","Data":"365bb82dbca333a1c0be98734a242477ddd0290bb75bb639cfbc566d1dd8b5d5"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.274535 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jgdv2" event={"ID":"abac02be-a75e-4ba2-ac00-4cb338681fec","Type":"ContainerStarted","Data":"82bec727f0b77e70fa4b0c1dbf45ce2224a65898abf8aa8341916715eb5a5aff"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.276808 4847 generic.go:334] "Generic (PLEG): container finished" podID="8b0fdd70-30cd-49a4-8072-044ef1cfd27a" containerID="358cdbf59419eac3710fd1b5c8c00a928ab6b639e9132906ac3d51c6355ddfd9" exitCode=0 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.276900 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a9e3-account-create-update-8s2rf" event={"ID":"8b0fdd70-30cd-49a4-8072-044ef1cfd27a","Type":"ContainerDied","Data":"358cdbf59419eac3710fd1b5c8c00a928ab6b639e9132906ac3d51c6355ddfd9"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.276949 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a9e3-account-create-update-8s2rf" event={"ID":"8b0fdd70-30cd-49a4-8072-044ef1cfd27a","Type":"ContainerStarted","Data":"c88cd6a2ef793d9c95c31d107d21c6406512e51182224887ad0839071c6aacbc"} Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.278411 4847 generic.go:334] "Generic (PLEG): container finished" podID="06fc695e-403f-4858-adfc-0ed1f8d414c7" containerID="a3da6562e66727d6f12a22f77c2c1e7013586e298df7b61811b075d98ad79966" exitCode=0 Dec 10 14:41:57 crc kubenswrapper[4847]: I1210 14:41:57.278497 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sj7rt" event={"ID":"06fc695e-403f-4858-adfc-0ed1f8d414c7","Type":"ContainerDied","Data":"a3da6562e66727d6f12a22f77c2c1e7013586e298df7b61811b075d98ad79966"} Dec 10 14:41:58 crc kubenswrapper[4847]: I1210 14:41:58.305831 4847 generic.go:334] "Generic (PLEG): container finished" podID="17e3cce4-5609-460e-b5b1-8e8a17962bc9" containerID="2fa46b3bcdfe3f7db9c04e77b2a3d8fb6d1a72530ce7e1a9e77b3ff1348e3d66" exitCode=0 Dec 10 14:41:58 crc kubenswrapper[4847]: I1210 14:41:58.306019 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-260c-account-create-update-cnfqw" event={"ID":"17e3cce4-5609-460e-b5b1-8e8a17962bc9","Type":"ContainerDied","Data":"2fa46b3bcdfe3f7db9c04e77b2a3d8fb6d1a72530ce7e1a9e77b3ff1348e3d66"} Dec 10 14:41:58 crc kubenswrapper[4847]: I1210 14:41:58.907617 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.039133 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts\") pod \"8711478e-be68-44d2-9471-225c0008b966\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.039206 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k59m\" (UniqueName: \"kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m\") pod \"8711478e-be68-44d2-9471-225c0008b966\" (UID: \"8711478e-be68-44d2-9471-225c0008b966\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.039985 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8711478e-be68-44d2-9471-225c0008b966" (UID: "8711478e-be68-44d2-9471-225c0008b966"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.045544 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m" (OuterVolumeSpecName: "kube-api-access-2k59m") pod "8711478e-be68-44d2-9471-225c0008b966" (UID: "8711478e-be68-44d2-9471-225c0008b966"). InnerVolumeSpecName "kube-api-access-2k59m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.046431 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.085044 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.092864 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.110000 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.140877 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts\") pod \"abac02be-a75e-4ba2-ac00-4cb338681fec\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141068 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65\") pod \"995e809e-0a8d-4b1e-acd8-b64571571651\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141118 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts\") pod \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141148 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2v9t\" (UniqueName: \"kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t\") pod \"abac02be-a75e-4ba2-ac00-4cb338681fec\" (UID: \"abac02be-a75e-4ba2-ac00-4cb338681fec\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141181 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q2lt\" (UniqueName: \"kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt\") pod \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\" (UID: \"8b0fdd70-30cd-49a4-8072-044ef1cfd27a\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141267 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts\") pod \"995e809e-0a8d-4b1e-acd8-b64571571651\" (UID: \"995e809e-0a8d-4b1e-acd8-b64571571651\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b0fdd70-30cd-49a4-8072-044ef1cfd27a" (UID: "8b0fdd70-30cd-49a4-8072-044ef1cfd27a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.141687 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8711478e-be68-44d2-9471-225c0008b966-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.142069 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k59m\" (UniqueName: \"kubernetes.io/projected/8711478e-be68-44d2-9471-225c0008b966-kube-api-access-2k59m\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.142078 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "995e809e-0a8d-4b1e-acd8-b64571571651" (UID: "995e809e-0a8d-4b1e-acd8-b64571571651"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.142261 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "abac02be-a75e-4ba2-ac00-4cb338681fec" (UID: "abac02be-a75e-4ba2-ac00-4cb338681fec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.145546 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt" (OuterVolumeSpecName: "kube-api-access-6q2lt") pod "8b0fdd70-30cd-49a4-8072-044ef1cfd27a" (UID: "8b0fdd70-30cd-49a4-8072-044ef1cfd27a"). InnerVolumeSpecName "kube-api-access-6q2lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.146257 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t" (OuterVolumeSpecName: "kube-api-access-t2v9t") pod "abac02be-a75e-4ba2-ac00-4cb338681fec" (UID: "abac02be-a75e-4ba2-ac00-4cb338681fec"). InnerVolumeSpecName "kube-api-access-t2v9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.147247 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65" (OuterVolumeSpecName: "kube-api-access-2dc65") pod "995e809e-0a8d-4b1e-acd8-b64571571651" (UID: "995e809e-0a8d-4b1e-acd8-b64571571651"). InnerVolumeSpecName "kube-api-access-2dc65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.243390 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts\") pod \"06fc695e-403f-4858-adfc-0ed1f8d414c7\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.243468 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74bqb\" (UniqueName: \"kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb\") pod \"06fc695e-403f-4858-adfc-0ed1f8d414c7\" (UID: \"06fc695e-403f-4858-adfc-0ed1f8d414c7\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.243884 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06fc695e-403f-4858-adfc-0ed1f8d414c7" (UID: "06fc695e-403f-4858-adfc-0ed1f8d414c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.243965 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/995e809e-0a8d-4b1e-acd8-b64571571651-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.244005 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abac02be-a75e-4ba2-ac00-4cb338681fec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.244015 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dc65\" (UniqueName: \"kubernetes.io/projected/995e809e-0a8d-4b1e-acd8-b64571571651-kube-api-access-2dc65\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.244028 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.244036 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2v9t\" (UniqueName: \"kubernetes.io/projected/abac02be-a75e-4ba2-ac00-4cb338681fec-kube-api-access-t2v9t\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.244046 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q2lt\" (UniqueName: \"kubernetes.io/projected/8b0fdd70-30cd-49a4-8072-044ef1cfd27a-kube-api-access-6q2lt\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.248224 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb" (OuterVolumeSpecName: "kube-api-access-74bqb") pod "06fc695e-403f-4858-adfc-0ed1f8d414c7" (UID: "06fc695e-403f-4858-adfc-0ed1f8d414c7"). InnerVolumeSpecName "kube-api-access-74bqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.315836 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2595-account-create-update-hxnfg" event={"ID":"995e809e-0a8d-4b1e-acd8-b64571571651","Type":"ContainerDied","Data":"0768579e60569b8539bf8c1be9edbb431a4233f7d9501466c3948d25d72c71fa"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.315872 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0768579e60569b8539bf8c1be9edbb431a4233f7d9501466c3948d25d72c71fa" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.315928 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2595-account-create-update-hxnfg" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.318530 4847 generic.go:334] "Generic (PLEG): container finished" podID="daaeae82-74d4-4d18-8178-58ca7246b605" containerID="b014b0d3566ba7ce4bf058ed5832d3c4f04c9dac5047c94ebf7cdc0492c05c90" exitCode=0 Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.318656 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hkv2r" event={"ID":"daaeae82-74d4-4d18-8178-58ca7246b605","Type":"ContainerDied","Data":"b014b0d3566ba7ce4bf058ed5832d3c4f04c9dac5047c94ebf7cdc0492c05c90"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.320594 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2l5wx" event={"ID":"8711478e-be68-44d2-9471-225c0008b966","Type":"ContainerDied","Data":"c73b68c4955b64b204968dcc29c52465151912ad9c66d27dcb74940aad1b07e3"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.320630 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c73b68c4955b64b204968dcc29c52465151912ad9c66d27dcb74940aad1b07e3" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.320689 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2l5wx" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.324483 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jgdv2" event={"ID":"abac02be-a75e-4ba2-ac00-4cb338681fec","Type":"ContainerDied","Data":"82bec727f0b77e70fa4b0c1dbf45ce2224a65898abf8aa8341916715eb5a5aff"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.324514 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82bec727f0b77e70fa4b0c1dbf45ce2224a65898abf8aa8341916715eb5a5aff" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.324528 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jgdv2" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.325726 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a9e3-account-create-update-8s2rf" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.325700 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a9e3-account-create-update-8s2rf" event={"ID":"8b0fdd70-30cd-49a4-8072-044ef1cfd27a","Type":"ContainerDied","Data":"c88cd6a2ef793d9c95c31d107d21c6406512e51182224887ad0839071c6aacbc"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.325995 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88cd6a2ef793d9c95c31d107d21c6406512e51182224887ad0839071c6aacbc" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.328893 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-sj7rt" event={"ID":"06fc695e-403f-4858-adfc-0ed1f8d414c7","Type":"ContainerDied","Data":"094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41"} Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.328932 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="094f95efda702aff9d1f6e7cc157798568e4de0099f4dcdcb323a4f61f8a6b41" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.328899 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-sj7rt" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.345250 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fc695e-403f-4858-adfc-0ed1f8d414c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.345279 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74bqb\" (UniqueName: \"kubernetes.io/projected/06fc695e-403f-4858-adfc-0ed1f8d414c7-kube-api-access-74bqb\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.575646 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.649346 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz475\" (UniqueName: \"kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475\") pod \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.649405 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts\") pod \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\" (UID: \"17e3cce4-5609-460e-b5b1-8e8a17962bc9\") " Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.649938 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17e3cce4-5609-460e-b5b1-8e8a17962bc9" (UID: "17e3cce4-5609-460e-b5b1-8e8a17962bc9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.653096 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475" (OuterVolumeSpecName: "kube-api-access-xz475") pod "17e3cce4-5609-460e-b5b1-8e8a17962bc9" (UID: "17e3cce4-5609-460e-b5b1-8e8a17962bc9"). InnerVolumeSpecName "kube-api-access-xz475". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.752783 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz475\" (UniqueName: \"kubernetes.io/projected/17e3cce4-5609-460e-b5b1-8e8a17962bc9-kube-api-access-xz475\") on node \"crc\" DevicePath \"\"" Dec 10 14:41:59 crc kubenswrapper[4847]: I1210 14:41:59.752855 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17e3cce4-5609-460e-b5b1-8e8a17962bc9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.337212 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-260c-account-create-update-cnfqw" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.337236 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-260c-account-create-update-cnfqw" event={"ID":"17e3cce4-5609-460e-b5b1-8e8a17962bc9","Type":"ContainerDied","Data":"5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654"} Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.338032 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5123d8bcdd47bcc274b4797bf299c18e2eb094e6a4621f962edac8421c7ed654" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.676878 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769195 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769235 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769312 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769427 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxxrx\" (UniqueName: \"kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769463 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769491 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.769517 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle\") pod \"daaeae82-74d4-4d18-8178-58ca7246b605\" (UID: \"daaeae82-74d4-4d18-8178-58ca7246b605\") " Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.770054 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.770523 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.798865 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx" (OuterVolumeSpecName: "kube-api-access-cxxrx") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "kube-api-access-cxxrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.804076 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.806264 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.810586 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts" (OuterVolumeSpecName: "scripts") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.812046 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daaeae82-74d4-4d18-8178-58ca7246b605" (UID: "daaeae82-74d4-4d18-8178-58ca7246b605"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871495 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871536 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxxrx\" (UniqueName: \"kubernetes.io/projected/daaeae82-74d4-4d18-8178-58ca7246b605-kube-api-access-cxxrx\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871551 4847 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871566 4847 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daaeae82-74d4-4d18-8178-58ca7246b605-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871577 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871591 4847 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daaeae82-74d4-4d18-8178-58ca7246b605-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:00 crc kubenswrapper[4847]: I1210 14:42:00.871601 4847 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daaeae82-74d4-4d18-8178-58ca7246b605-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.011246 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.011325 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.207001 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-czhlz"] Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211267 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daaeae82-74d4-4d18-8178-58ca7246b605" containerName="swift-ring-rebalance" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211301 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="daaeae82-74d4-4d18-8178-58ca7246b605" containerName="swift-ring-rebalance" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211329 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fc695e-403f-4858-adfc-0ed1f8d414c7" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211338 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fc695e-403f-4858-adfc-0ed1f8d414c7" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211355 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995e809e-0a8d-4b1e-acd8-b64571571651" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211363 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="995e809e-0a8d-4b1e-acd8-b64571571651" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211400 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e3cce4-5609-460e-b5b1-8e8a17962bc9" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211408 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e3cce4-5609-460e-b5b1-8e8a17962bc9" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211418 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abac02be-a75e-4ba2-ac00-4cb338681fec" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211427 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="abac02be-a75e-4ba2-ac00-4cb338681fec" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211437 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8711478e-be68-44d2-9471-225c0008b966" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211444 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8711478e-be68-44d2-9471-225c0008b966" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: E1210 14:42:01.211456 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b0fdd70-30cd-49a4-8072-044ef1cfd27a" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211463 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b0fdd70-30cd-49a4-8072-044ef1cfd27a" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211888 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e3cce4-5609-460e-b5b1-8e8a17962bc9" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211930 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="995e809e-0a8d-4b1e-acd8-b64571571651" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211940 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b0fdd70-30cd-49a4-8072-044ef1cfd27a" containerName="mariadb-account-create-update" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211949 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fc695e-403f-4858-adfc-0ed1f8d414c7" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211958 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="abac02be-a75e-4ba2-ac00-4cb338681fec" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211972 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="daaeae82-74d4-4d18-8178-58ca7246b605" containerName="swift-ring-rebalance" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.211986 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8711478e-be68-44d2-9471-225c0008b966" containerName="mariadb-database-create" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.212549 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.214242 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.214321 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m498w" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.224214 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-czhlz"] Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.278020 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.278147 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.278203 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spcht\" (UniqueName: \"kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.278226 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.348708 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hkv2r" event={"ID":"daaeae82-74d4-4d18-8178-58ca7246b605","Type":"ContainerDied","Data":"213dd121d779bcd20af6b7c1c43cdaa2dc8743d4ad6b706bb0443ae7400d9939"} Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.348775 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="213dd121d779bcd20af6b7c1c43cdaa2dc8743d4ad6b706bb0443ae7400d9939" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.348863 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hkv2r" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.354154 4847 generic.go:334] "Generic (PLEG): container finished" podID="2d33f956-6feb-4c03-88f9-95185d25f781" containerID="a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05" exitCode=0 Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.354250 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerDied","Data":"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05"} Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.355812 4847 generic.go:334] "Generic (PLEG): container finished" podID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerID="415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16" exitCode=0 Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.355855 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerDied","Data":"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16"} Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.380257 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.380351 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.380373 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spcht\" (UniqueName: \"kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.380447 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.385038 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.386235 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.387783 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.403979 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spcht\" (UniqueName: \"kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht\") pod \"glance-db-sync-czhlz\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.571139 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czhlz" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.836811 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2xjh7" podUID="b962e42f-e21b-4016-9587-111fb6eb68ac" containerName="ovn-controller" probeResult="failure" output=< Dec 10 14:42:01 crc kubenswrapper[4847]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 14:42:01 crc kubenswrapper[4847]: > Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.921233 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:42:01 crc kubenswrapper[4847]: I1210 14:42:01.930473 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rnkkd" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.218674 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-czhlz"] Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.263918 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2xjh7-config-jpk96"] Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.264901 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.268857 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.275596 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7-config-jpk96"] Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301527 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301596 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301651 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301687 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301755 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46fj8\" (UniqueName: \"kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.301787 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.366062 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerStarted","Data":"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d"} Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.366519 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.368086 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerStarted","Data":"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683"} Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.368546 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.370056 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czhlz" event={"ID":"32400504-9741-4abf-bdbf-a8ea9fcb8e4e","Type":"ContainerStarted","Data":"7bb8148a26e0c12e9fe13cd8d060b6abc265a52bdaa0fd9fdff50a16f9d81218"} Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.391617 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.147391207 podStartE2EDuration="1m2.391586429s" podCreationTimestamp="2025-12-10 14:41:00 +0000 UTC" firstStartedPulling="2025-12-10 14:41:18.860326997 +0000 UTC m=+1028.429544627" lastFinishedPulling="2025-12-10 14:41:27.104522219 +0000 UTC m=+1036.673739849" observedRunningTime="2025-12-10 14:42:02.389690727 +0000 UTC m=+1071.958908357" watchObservedRunningTime="2025-12-10 14:42:02.391586429 +0000 UTC m=+1071.960804059" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403270 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403388 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46fj8\" (UniqueName: \"kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403453 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403526 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403671 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.403871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.404108 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.404875 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.404924 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.405052 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.406196 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.424259 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.342603756 podStartE2EDuration="1m2.424237231s" podCreationTimestamp="2025-12-10 14:41:00 +0000 UTC" firstStartedPulling="2025-12-10 14:41:18.261002416 +0000 UTC m=+1027.830220046" lastFinishedPulling="2025-12-10 14:41:27.342635891 +0000 UTC m=+1036.911853521" observedRunningTime="2025-12-10 14:42:02.415003415 +0000 UTC m=+1071.984221055" watchObservedRunningTime="2025-12-10 14:42:02.424237231 +0000 UTC m=+1071.993454861" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.425562 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46fj8\" (UniqueName: \"kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8\") pod \"ovn-controller-2xjh7-config-jpk96\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:02 crc kubenswrapper[4847]: I1210 14:42:02.585148 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:03 crc kubenswrapper[4847]: I1210 14:42:03.203024 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7-config-jpk96"] Dec 10 14:42:03 crc kubenswrapper[4847]: W1210 14:42:03.213103 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae38bc86_b150_4426_8ab5_2f66bd4c6e40.slice/crio-036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b WatchSource:0}: Error finding container 036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b: Status 404 returned error can't find the container with id 036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b Dec 10 14:42:03 crc kubenswrapper[4847]: I1210 14:42:03.382036 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-jpk96" event={"ID":"ae38bc86-b150-4426-8ab5-2f66bd4c6e40","Type":"ContainerStarted","Data":"036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b"} Dec 10 14:42:03 crc kubenswrapper[4847]: I1210 14:42:03.385867 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a4d0db06-c1bd-4af6-b78e-a1703d914c6d","Type":"ContainerStarted","Data":"04f335b4b5bf2a7bdd4054d5bc49857b622a24a3452c921353d8809865364470"} Dec 10 14:42:03 crc kubenswrapper[4847]: I1210 14:42:03.386856 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 10 14:42:03 crc kubenswrapper[4847]: I1210 14:42:03.422539 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.984638331 podStartE2EDuration="28.42245003s" podCreationTimestamp="2025-12-10 14:41:35 +0000 UTC" firstStartedPulling="2025-12-10 14:41:36.498095023 +0000 UTC m=+1046.067312653" lastFinishedPulling="2025-12-10 14:42:02.935906722 +0000 UTC m=+1072.505124352" observedRunningTime="2025-12-10 14:42:03.412171826 +0000 UTC m=+1072.981389456" watchObservedRunningTime="2025-12-10 14:42:03.42245003 +0000 UTC m=+1072.991667660" Dec 10 14:42:04 crc kubenswrapper[4847]: I1210 14:42:04.395557 4847 generic.go:334] "Generic (PLEG): container finished" podID="ae38bc86-b150-4426-8ab5-2f66bd4c6e40" containerID="a8df61f54bd8405abd20ac66610931007cf65b734093972a9a9ae7133c03d0ea" exitCode=0 Dec 10 14:42:04 crc kubenswrapper[4847]: I1210 14:42:04.395655 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-jpk96" event={"ID":"ae38bc86-b150-4426-8ab5-2f66bd4c6e40","Type":"ContainerDied","Data":"a8df61f54bd8405abd20ac66610931007cf65b734093972a9a9ae7133c03d0ea"} Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.743334 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868596 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868663 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868782 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run" (OuterVolumeSpecName: "var-run") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868864 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868875 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868891 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868913 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.868971 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.869042 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46fj8\" (UniqueName: \"kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8\") pod \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\" (UID: \"ae38bc86-b150-4426-8ab5-2f66bd4c6e40\") " Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870409 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts" (OuterVolumeSpecName: "scripts") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870644 4847 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870668 4847 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870679 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870687 4847 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.870817 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.910090 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8" (OuterVolumeSpecName: "kube-api-access-46fj8") pod "ae38bc86-b150-4426-8ab5-2f66bd4c6e40" (UID: "ae38bc86-b150-4426-8ab5-2f66bd4c6e40"). InnerVolumeSpecName "kube-api-access-46fj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.972292 4847 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:05 crc kubenswrapper[4847]: I1210 14:42:05.972345 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46fj8\" (UniqueName: \"kubernetes.io/projected/ae38bc86-b150-4426-8ab5-2f66bd4c6e40-kube-api-access-46fj8\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.516383 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-jpk96" event={"ID":"ae38bc86-b150-4426-8ab5-2f66bd4c6e40","Type":"ContainerDied","Data":"036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b"} Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.517053 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="036ed576f9bcef7223f3480fa3fe9917153f23222327b7de10bd73a355e4f50b" Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.516663 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-jpk96" Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.951567 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-2xjh7" Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.954120 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2xjh7-config-jpk96"] Dec 10 14:42:06 crc kubenswrapper[4847]: I1210 14:42:06.972311 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2xjh7-config-jpk96"] Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.031973 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2xjh7-config-b4n4n"] Dec 10 14:42:07 crc kubenswrapper[4847]: E1210 14:42:07.032346 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae38bc86-b150-4426-8ab5-2f66bd4c6e40" containerName="ovn-config" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.032363 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae38bc86-b150-4426-8ab5-2f66bd4c6e40" containerName="ovn-config" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.032527 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae38bc86-b150-4426-8ab5-2f66bd4c6e40" containerName="ovn-config" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.033070 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.039915 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.045379 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7-config-b4n4n"] Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137498 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137554 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137614 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9h27\" (UniqueName: \"kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137644 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137669 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.137746 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.239737 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.239800 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.239899 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9h27\" (UniqueName: \"kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.239929 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.239959 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.240021 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.240451 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.243194 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.243266 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.243588 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.244046 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.265332 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9h27\" (UniqueName: \"kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27\") pod \"ovn-controller-2xjh7-config-b4n4n\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.356425 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:07 crc kubenswrapper[4847]: I1210 14:42:07.950270 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2xjh7-config-b4n4n"] Dec 10 14:42:08 crc kubenswrapper[4847]: I1210 14:42:08.533109 4847 generic.go:334] "Generic (PLEG): container finished" podID="e5b75e08-4a3d-410f-9c10-81f695250989" containerID="68ea8b37c6643ca1fc7703f6f8985f1cf33536eeab82c06721873fa242986cb4" exitCode=0 Dec 10 14:42:08 crc kubenswrapper[4847]: I1210 14:42:08.533258 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-b4n4n" event={"ID":"e5b75e08-4a3d-410f-9c10-81f695250989","Type":"ContainerDied","Data":"68ea8b37c6643ca1fc7703f6f8985f1cf33536eeab82c06721873fa242986cb4"} Dec 10 14:42:08 crc kubenswrapper[4847]: I1210 14:42:08.533400 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-b4n4n" event={"ID":"e5b75e08-4a3d-410f-9c10-81f695250989","Type":"ContainerStarted","Data":"c56bcd2cbc8581da2981fc7e7e264f12fa69c9dd934d644e4738931d9a7c502d"} Dec 10 14:42:08 crc kubenswrapper[4847]: I1210 14:42:08.768129 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae38bc86-b150-4426-8ab5-2f66bd4c6e40" path="/var/lib/kubelet/pods/ae38bc86-b150-4426-8ab5-2f66bd4c6e40/volumes" Dec 10 14:42:10 crc kubenswrapper[4847]: I1210 14:42:10.772732 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:42:10 crc kubenswrapper[4847]: I1210 14:42:10.795411 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/88c96cdf-8002-4829-9f8c-3abfc8315722-etc-swift\") pod \"swift-storage-0\" (UID: \"88c96cdf-8002-4829-9f8c-3abfc8315722\") " pod="openstack/swift-storage-0" Dec 10 14:42:10 crc kubenswrapper[4847]: I1210 14:42:10.953856 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.246025 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.299159 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.691245 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bb6jt"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.692479 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.708040 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bb6jt"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.718262 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.718315 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7tz5\" (UniqueName: \"kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.720055 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bd7a-account-create-update-l2vzd"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.721113 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.724345 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.734951 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bd7a-account-create-update-l2vzd"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.793751 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-cc2e-account-create-update-qz9b9"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.795071 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.798219 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.803915 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-fbd72"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.807611 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.819949 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820024 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7lcr\" (UniqueName: \"kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820056 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820104 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7tz5\" (UniqueName: \"kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820142 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn6qt\" (UniqueName: \"kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820204 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drggl\" (UniqueName: \"kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820242 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.820273 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.821073 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.821786 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cc2e-account-create-update-qz9b9"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.834549 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fbd72"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.879563 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7tz5\" (UniqueName: \"kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5\") pod \"cinder-db-create-bb6jt\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922371 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922422 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922499 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7lcr\" (UniqueName: \"kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922523 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922542 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn6qt\" (UniqueName: \"kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.922582 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drggl\" (UniqueName: \"kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.923822 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.924325 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.924654 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.940589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drggl\" (UniqueName: \"kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl\") pod \"barbican-db-create-fbd72\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.941339 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7lcr\" (UniqueName: \"kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr\") pod \"barbican-cc2e-account-create-update-qz9b9\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.946794 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn6qt\" (UniqueName: \"kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt\") pod \"cinder-bd7a-account-create-update-l2vzd\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.994589 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-4cxxl"] Dec 10 14:42:12 crc kubenswrapper[4847]: I1210 14:42:12.997285 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.009000 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4cxxl"] Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.027056 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.055075 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.082781 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ssqhj"] Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.087801 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.091980 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.091998 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.092288 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.092382 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mz28n" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.094948 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ssqhj"] Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.116603 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-906f-account-create-update-xk8nx"] Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.116972 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.118821 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.121853 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.130165 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.130214 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x56tt\" (UniqueName: \"kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.131501 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.138760 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-906f-account-create-update-xk8nx"] Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.231763 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x56tt\" (UniqueName: \"kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.231834 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbjz4\" (UniqueName: \"kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.231873 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.232002 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg7tg\" (UniqueName: \"kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.232044 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.232072 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.232096 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.233024 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.248595 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x56tt\" (UniqueName: \"kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt\") pod \"neutron-db-create-4cxxl\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.333352 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg7tg\" (UniqueName: \"kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.333427 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.333456 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.333508 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbjz4\" (UniqueName: \"kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.333559 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.334324 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.338593 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.339992 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.341619 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.359066 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg7tg\" (UniqueName: \"kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg\") pod \"neutron-906f-account-create-update-xk8nx\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.362604 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbjz4\" (UniqueName: \"kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4\") pod \"keystone-db-sync-ssqhj\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.424077 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:13 crc kubenswrapper[4847]: I1210 14:42:13.435190 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:16 crc kubenswrapper[4847]: I1210 14:42:16.070523 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 10 14:42:18 crc kubenswrapper[4847]: E1210 14:42:18.421878 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 10 14:42:18 crc kubenswrapper[4847]: E1210 14:42:18.422409 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-spcht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-czhlz_openstack(32400504-9741-4abf-bdbf-a8ea9fcb8e4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:42:18 crc kubenswrapper[4847]: E1210 14:42:18.423768 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-czhlz" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.554574 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.695270 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2xjh7-config-b4n4n" event={"ID":"e5b75e08-4a3d-410f-9c10-81f695250989","Type":"ContainerDied","Data":"c56bcd2cbc8581da2981fc7e7e264f12fa69c9dd934d644e4738931d9a7c502d"} Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.695616 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c56bcd2cbc8581da2981fc7e7e264f12fa69c9dd934d644e4738931d9a7c502d" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.695293 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2xjh7-config-b4n4n" Dec 10 14:42:18 crc kubenswrapper[4847]: E1210 14:42:18.700245 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-czhlz" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729436 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729567 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9h27\" (UniqueName: \"kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729602 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729663 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729724 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.729778 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn\") pod \"e5b75e08-4a3d-410f-9c10-81f695250989\" (UID: \"e5b75e08-4a3d-410f-9c10-81f695250989\") " Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.730176 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.730218 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.730812 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.730849 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run" (OuterVolumeSpecName: "var-run") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.731696 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts" (OuterVolumeSpecName: "scripts") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.742348 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27" (OuterVolumeSpecName: "kube-api-access-k9h27") pod "e5b75e08-4a3d-410f-9c10-81f695250989" (UID: "e5b75e08-4a3d-410f-9c10-81f695250989"). InnerVolumeSpecName "kube-api-access-k9h27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842812 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9h27\" (UniqueName: \"kubernetes.io/projected/e5b75e08-4a3d-410f-9c10-81f695250989-kube-api-access-k9h27\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842845 4847 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842858 4847 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842869 4847 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842881 4847 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e5b75e08-4a3d-410f-9c10-81f695250989-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.842892 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5b75e08-4a3d-410f-9c10-81f695250989-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:18 crc kubenswrapper[4847]: I1210 14:42:18.897260 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fbd72"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.136840 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bb6jt"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.146029 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-906f-account-create-update-xk8nx"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.193224 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.301143 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ssqhj"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.320974 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-cc2e-account-create-update-qz9b9"] Dec 10 14:42:19 crc kubenswrapper[4847]: W1210 14:42:19.327459 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda11592b8_0374_4b1d_a579_64b72830148a.slice/crio-0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a WatchSource:0}: Error finding container 0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a: Status 404 returned error can't find the container with id 0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.333050 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-4cxxl"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.339224 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bd7a-account-create-update-l2vzd"] Dec 10 14:42:19 crc kubenswrapper[4847]: W1210 14:42:19.346876 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bba7667_cbb4_4636_9197_2409e86b8cde.slice/crio-fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b WatchSource:0}: Error finding container fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b: Status 404 returned error can't find the container with id fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.630262 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2xjh7-config-b4n4n"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.642288 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2xjh7-config-b4n4n"] Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.706171 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bd7a-account-create-update-l2vzd" event={"ID":"0bba7667-cbb4-4636-9197-2409e86b8cde","Type":"ContainerStarted","Data":"8d58579c90642bdd77f103403ec3a5e487b8e23736d073c9ee80a295d4f82afe"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.706223 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bd7a-account-create-update-l2vzd" event={"ID":"0bba7667-cbb4-4636-9197-2409e86b8cde","Type":"ContainerStarted","Data":"fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.708661 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cc2e-account-create-update-qz9b9" event={"ID":"a11592b8-0374-4b1d-a579-64b72830148a","Type":"ContainerStarted","Data":"84a54dec72b3da41bc66d1467dc78819f13c2500cbc38621a3e4180fd8e79163"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.708708 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cc2e-account-create-update-qz9b9" event={"ID":"a11592b8-0374-4b1d-a579-64b72830148a","Type":"ContainerStarted","Data":"0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.710647 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"726e838b1ac85090a7742e2ddc81c66ed1e3ffc13a91d344e2d50810b254e72d"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.712848 4847 generic.go:334] "Generic (PLEG): container finished" podID="17bbab02-b41c-4b58-a235-0c88b7eeb1b3" containerID="99bd80ca3b3f97d20133979a88bec62c4a665c1b6ca7f05b1da592b91b6e6d6a" exitCode=0 Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.712963 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fbd72" event={"ID":"17bbab02-b41c-4b58-a235-0c88b7eeb1b3","Type":"ContainerDied","Data":"99bd80ca3b3f97d20133979a88bec62c4a665c1b6ca7f05b1da592b91b6e6d6a"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.713050 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fbd72" event={"ID":"17bbab02-b41c-4b58-a235-0c88b7eeb1b3","Type":"ContainerStarted","Data":"f6d40dfa9954d1be1b851667c172687d5ee0f27cf37e97803258374ebd9d3cad"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.715932 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-906f-account-create-update-xk8nx" event={"ID":"e5096792-9e2e-4843-85da-9e14112d056a","Type":"ContainerStarted","Data":"c44e9fc63261d85023464a8913b4fbe3f982017c538e9dd12263d29e039c4a61"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.716113 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-906f-account-create-update-xk8nx" event={"ID":"e5096792-9e2e-4843-85da-9e14112d056a","Type":"ContainerStarted","Data":"f6eeda10a5a7c0b007403180e5858bf7f77738ef2774af162f96acfe9eaeac48"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.719673 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bb6jt" event={"ID":"7148ddb7-2803-41ab-bf95-f820c60614d1","Type":"ContainerStarted","Data":"ecfca552d6bf41de43d26684aab0607b227fc027286c65be3ae47cce6d2b7a97"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.719723 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bb6jt" event={"ID":"7148ddb7-2803-41ab-bf95-f820c60614d1","Type":"ContainerStarted","Data":"0a19e130508e890cd38533a457660a00a57b4c227c9402f673a32fc77cae4a2f"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.724087 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4cxxl" event={"ID":"3b75aa68-b977-4614-877b-9828d3045226","Type":"ContainerStarted","Data":"80ae245596e55e80cde46d6afdd39ba3f0991e25a7959f972b9d4db4507ebafc"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.724154 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4cxxl" event={"ID":"3b75aa68-b977-4614-877b-9828d3045226","Type":"ContainerStarted","Data":"a4d6e90cb273e580d5246774125ca4ad8cceb9c0d583a58c31da00fe9abd45de"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.727845 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ssqhj" event={"ID":"ecf994d1-33cd-404a-bec2-01956b20b78d","Type":"ContainerStarted","Data":"fb9ac85c1cffaa567101621a05ce9353dc06fd2ce925a7fc676e25c5589a60ef"} Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.731316 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-bd7a-account-create-update-l2vzd" podStartSLOduration=7.731275591 podStartE2EDuration="7.731275591s" podCreationTimestamp="2025-12-10 14:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:19.721975473 +0000 UTC m=+1089.291193103" watchObservedRunningTime="2025-12-10 14:42:19.731275591 +0000 UTC m=+1089.300493221" Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.743963 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-bb6jt" podStartSLOduration=7.74394299 podStartE2EDuration="7.74394299s" podCreationTimestamp="2025-12-10 14:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:19.736904366 +0000 UTC m=+1089.306121996" watchObservedRunningTime="2025-12-10 14:42:19.74394299 +0000 UTC m=+1089.313160620" Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.777847 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-cc2e-account-create-update-qz9b9" podStartSLOduration=7.777823015 podStartE2EDuration="7.777823015s" podCreationTimestamp="2025-12-10 14:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:19.751769086 +0000 UTC m=+1089.320986726" watchObservedRunningTime="2025-12-10 14:42:19.777823015 +0000 UTC m=+1089.347040645" Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.787666 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-906f-account-create-update-xk8nx" podStartSLOduration=6.787643446 podStartE2EDuration="6.787643446s" podCreationTimestamp="2025-12-10 14:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:19.781654821 +0000 UTC m=+1089.350872451" watchObservedRunningTime="2025-12-10 14:42:19.787643446 +0000 UTC m=+1089.356861076" Dec 10 14:42:19 crc kubenswrapper[4847]: I1210 14:42:19.795489 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-4cxxl" podStartSLOduration=7.795469141 podStartE2EDuration="7.795469141s" podCreationTimestamp="2025-12-10 14:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:19.794359231 +0000 UTC m=+1089.363576871" watchObservedRunningTime="2025-12-10 14:42:19.795469141 +0000 UTC m=+1089.364686771" Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.737506 4847 generic.go:334] "Generic (PLEG): container finished" podID="7148ddb7-2803-41ab-bf95-f820c60614d1" containerID="ecfca552d6bf41de43d26684aab0607b227fc027286c65be3ae47cce6d2b7a97" exitCode=0 Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.737592 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bb6jt" event={"ID":"7148ddb7-2803-41ab-bf95-f820c60614d1","Type":"ContainerDied","Data":"ecfca552d6bf41de43d26684aab0607b227fc027286c65be3ae47cce6d2b7a97"} Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.738954 4847 generic.go:334] "Generic (PLEG): container finished" podID="3b75aa68-b977-4614-877b-9828d3045226" containerID="80ae245596e55e80cde46d6afdd39ba3f0991e25a7959f972b9d4db4507ebafc" exitCode=0 Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.739012 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4cxxl" event={"ID":"3b75aa68-b977-4614-877b-9828d3045226","Type":"ContainerDied","Data":"80ae245596e55e80cde46d6afdd39ba3f0991e25a7959f972b9d4db4507ebafc"} Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.742673 4847 generic.go:334] "Generic (PLEG): container finished" podID="0bba7667-cbb4-4636-9197-2409e86b8cde" containerID="8d58579c90642bdd77f103403ec3a5e487b8e23736d073c9ee80a295d4f82afe" exitCode=0 Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.742770 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bd7a-account-create-update-l2vzd" event={"ID":"0bba7667-cbb4-4636-9197-2409e86b8cde","Type":"ContainerDied","Data":"8d58579c90642bdd77f103403ec3a5e487b8e23736d073c9ee80a295d4f82afe"} Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.744096 4847 generic.go:334] "Generic (PLEG): container finished" podID="a11592b8-0374-4b1d-a579-64b72830148a" containerID="84a54dec72b3da41bc66d1467dc78819f13c2500cbc38621a3e4180fd8e79163" exitCode=0 Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.744156 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cc2e-account-create-update-qz9b9" event={"ID":"a11592b8-0374-4b1d-a579-64b72830148a","Type":"ContainerDied","Data":"84a54dec72b3da41bc66d1467dc78819f13c2500cbc38621a3e4180fd8e79163"} Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.745396 4847 generic.go:334] "Generic (PLEG): container finished" podID="e5096792-9e2e-4843-85da-9e14112d056a" containerID="c44e9fc63261d85023464a8913b4fbe3f982017c538e9dd12263d29e039c4a61" exitCode=0 Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.745440 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-906f-account-create-update-xk8nx" event={"ID":"e5096792-9e2e-4843-85da-9e14112d056a","Type":"ContainerDied","Data":"c44e9fc63261d85023464a8913b4fbe3f982017c538e9dd12263d29e039c4a61"} Dec 10 14:42:20 crc kubenswrapper[4847]: I1210 14:42:20.780589 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b75e08-4a3d-410f-9c10-81f695250989" path="/var/lib/kubelet/pods/e5b75e08-4a3d-410f-9c10-81f695250989/volumes" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.093019 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.135492 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts\") pod \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.135633 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drggl\" (UniqueName: \"kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl\") pod \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\" (UID: \"17bbab02-b41c-4b58-a235-0c88b7eeb1b3\") " Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.136106 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17bbab02-b41c-4b58-a235-0c88b7eeb1b3" (UID: "17bbab02-b41c-4b58-a235-0c88b7eeb1b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.141947 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl" (OuterVolumeSpecName: "kube-api-access-drggl") pod "17bbab02-b41c-4b58-a235-0c88b7eeb1b3" (UID: "17bbab02-b41c-4b58-a235-0c88b7eeb1b3"). InnerVolumeSpecName "kube-api-access-drggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.237481 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.237523 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drggl\" (UniqueName: \"kubernetes.io/projected/17bbab02-b41c-4b58-a235-0c88b7eeb1b3-kube-api-access-drggl\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.755647 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fbd72" event={"ID":"17bbab02-b41c-4b58-a235-0c88b7eeb1b3","Type":"ContainerDied","Data":"f6d40dfa9954d1be1b851667c172687d5ee0f27cf37e97803258374ebd9d3cad"} Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.755691 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6d40dfa9954d1be1b851667c172687d5ee0f27cf37e97803258374ebd9d3cad" Dec 10 14:42:21 crc kubenswrapper[4847]: I1210 14:42:21.755941 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fbd72" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.094109 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.156374 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts\") pod \"e5096792-9e2e-4843-85da-9e14112d056a\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.156456 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg7tg\" (UniqueName: \"kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg\") pod \"e5096792-9e2e-4843-85da-9e14112d056a\" (UID: \"e5096792-9e2e-4843-85da-9e14112d056a\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.157256 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5096792-9e2e-4843-85da-9e14112d056a" (UID: "e5096792-9e2e-4843-85da-9e14112d056a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.177084 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg" (OuterVolumeSpecName: "kube-api-access-lg7tg") pod "e5096792-9e2e-4843-85da-9e14112d056a" (UID: "e5096792-9e2e-4843-85da-9e14112d056a"). InnerVolumeSpecName "kube-api-access-lg7tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.257986 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5096792-9e2e-4843-85da-9e14112d056a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.258022 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg7tg\" (UniqueName: \"kubernetes.io/projected/e5096792-9e2e-4843-85da-9e14112d056a-kube-api-access-lg7tg\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.280000 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.284656 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.313974 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.316810 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.360914 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7tz5\" (UniqueName: \"kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5\") pod \"7148ddb7-2803-41ab-bf95-f820c60614d1\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361004 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts\") pod \"a11592b8-0374-4b1d-a579-64b72830148a\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361061 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts\") pod \"3b75aa68-b977-4614-877b-9828d3045226\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361088 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x56tt\" (UniqueName: \"kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt\") pod \"3b75aa68-b977-4614-877b-9828d3045226\" (UID: \"3b75aa68-b977-4614-877b-9828d3045226\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361228 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts\") pod \"0bba7667-cbb4-4636-9197-2409e86b8cde\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361388 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts\") pod \"7148ddb7-2803-41ab-bf95-f820c60614d1\" (UID: \"7148ddb7-2803-41ab-bf95-f820c60614d1\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361430 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn6qt\" (UniqueName: \"kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt\") pod \"0bba7667-cbb4-4636-9197-2409e86b8cde\" (UID: \"0bba7667-cbb4-4636-9197-2409e86b8cde\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.361468 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7lcr\" (UniqueName: \"kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr\") pod \"a11592b8-0374-4b1d-a579-64b72830148a\" (UID: \"a11592b8-0374-4b1d-a579-64b72830148a\") " Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.369737 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0bba7667-cbb4-4636-9197-2409e86b8cde" (UID: "0bba7667-cbb4-4636-9197-2409e86b8cde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.370250 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7148ddb7-2803-41ab-bf95-f820c60614d1" (UID: "7148ddb7-2803-41ab-bf95-f820c60614d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.372703 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt" (OuterVolumeSpecName: "kube-api-access-x56tt") pod "3b75aa68-b977-4614-877b-9828d3045226" (UID: "3b75aa68-b977-4614-877b-9828d3045226"). InnerVolumeSpecName "kube-api-access-x56tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.373274 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a11592b8-0374-4b1d-a579-64b72830148a" (UID: "a11592b8-0374-4b1d-a579-64b72830148a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.373763 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3b75aa68-b977-4614-877b-9828d3045226" (UID: "3b75aa68-b977-4614-877b-9828d3045226"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.376649 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr" (OuterVolumeSpecName: "kube-api-access-k7lcr") pod "a11592b8-0374-4b1d-a579-64b72830148a" (UID: "a11592b8-0374-4b1d-a579-64b72830148a"). InnerVolumeSpecName "kube-api-access-k7lcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.379556 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5" (OuterVolumeSpecName: "kube-api-access-g7tz5") pod "7148ddb7-2803-41ab-bf95-f820c60614d1" (UID: "7148ddb7-2803-41ab-bf95-f820c60614d1"). InnerVolumeSpecName "kube-api-access-g7tz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.386851 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt" (OuterVolumeSpecName: "kube-api-access-zn6qt") pod "0bba7667-cbb4-4636-9197-2409e86b8cde" (UID: "0bba7667-cbb4-4636-9197-2409e86b8cde"). InnerVolumeSpecName "kube-api-access-zn6qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463205 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b75aa68-b977-4614-877b-9828d3045226-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463239 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x56tt\" (UniqueName: \"kubernetes.io/projected/3b75aa68-b977-4614-877b-9828d3045226-kube-api-access-x56tt\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463253 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0bba7667-cbb4-4636-9197-2409e86b8cde-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463266 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7148ddb7-2803-41ab-bf95-f820c60614d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463278 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn6qt\" (UniqueName: \"kubernetes.io/projected/0bba7667-cbb4-4636-9197-2409e86b8cde-kube-api-access-zn6qt\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463290 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7lcr\" (UniqueName: \"kubernetes.io/projected/a11592b8-0374-4b1d-a579-64b72830148a-kube-api-access-k7lcr\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463302 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7tz5\" (UniqueName: \"kubernetes.io/projected/7148ddb7-2803-41ab-bf95-f820c60614d1-kube-api-access-g7tz5\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.463311 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a11592b8-0374-4b1d-a579-64b72830148a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.766070 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-cc2e-account-create-update-qz9b9" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.770647 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-906f-account-create-update-xk8nx" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.778829 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-cc2e-account-create-update-qz9b9" event={"ID":"a11592b8-0374-4b1d-a579-64b72830148a","Type":"ContainerDied","Data":"0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a"} Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.778873 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c3722ac2c2908ba492088bb950f5b0803737fabd3586224dbaaf654e33cdc8a" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.778886 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-906f-account-create-update-xk8nx" event={"ID":"e5096792-9e2e-4843-85da-9e14112d056a","Type":"ContainerDied","Data":"f6eeda10a5a7c0b007403180e5858bf7f77738ef2774af162f96acfe9eaeac48"} Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.778897 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6eeda10a5a7c0b007403180e5858bf7f77738ef2774af162f96acfe9eaeac48" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.781809 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bb6jt" event={"ID":"7148ddb7-2803-41ab-bf95-f820c60614d1","Type":"ContainerDied","Data":"0a19e130508e890cd38533a457660a00a57b4c227c9402f673a32fc77cae4a2f"} Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.781868 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a19e130508e890cd38533a457660a00a57b4c227c9402f673a32fc77cae4a2f" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.781925 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bb6jt" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.788932 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-4cxxl" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.788929 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-4cxxl" event={"ID":"3b75aa68-b977-4614-877b-9828d3045226","Type":"ContainerDied","Data":"a4d6e90cb273e580d5246774125ca4ad8cceb9c0d583a58c31da00fe9abd45de"} Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.789037 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4d6e90cb273e580d5246774125ca4ad8cceb9c0d583a58c31da00fe9abd45de" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.790462 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bd7a-account-create-update-l2vzd" event={"ID":"0bba7667-cbb4-4636-9197-2409e86b8cde","Type":"ContainerDied","Data":"fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b"} Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.790492 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc97b76cca0910b12d761224059246826e65c43e1e32bf0aaeb58594a10ab04b" Dec 10 14:42:22 crc kubenswrapper[4847]: I1210 14:42:22.790538 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bd7a-account-create-update-l2vzd" Dec 10 14:42:26 crc kubenswrapper[4847]: I1210 14:42:26.927441 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ssqhj" event={"ID":"ecf994d1-33cd-404a-bec2-01956b20b78d","Type":"ContainerStarted","Data":"8347dbf81b888a31f98854a1ff7e74c2100b2619c2a2325b51d66e1270c79abf"} Dec 10 14:42:27 crc kubenswrapper[4847]: I1210 14:42:26.946268 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"9b9975be589fe81e5136cd73862e74ca709bddad076027bf0898c793126f9d5c"} Dec 10 14:42:27 crc kubenswrapper[4847]: I1210 14:42:26.946309 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"91919454e7955563e26c723f1c9d9146c292630930acef0993876b6774d14244"} Dec 10 14:42:27 crc kubenswrapper[4847]: I1210 14:42:26.946319 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"0e8c4da21c7f3ad0dd13185df6810a4c545292c42f070056c86105b05348ea55"} Dec 10 14:42:27 crc kubenswrapper[4847]: I1210 14:42:26.961158 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ssqhj" podStartSLOduration=7.319594997 podStartE2EDuration="13.961138658s" podCreationTimestamp="2025-12-10 14:42:13 +0000 UTC" firstStartedPulling="2025-12-10 14:42:19.306354773 +0000 UTC m=+1088.875572413" lastFinishedPulling="2025-12-10 14:42:25.947898444 +0000 UTC m=+1095.517116074" observedRunningTime="2025-12-10 14:42:26.956046138 +0000 UTC m=+1096.525263768" watchObservedRunningTime="2025-12-10 14:42:26.961138658 +0000 UTC m=+1096.530356308" Dec 10 14:42:27 crc kubenswrapper[4847]: I1210 14:42:27.956871 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"d8466a10612a88f600180b49c3fdcc8152505309f60afb51800a6bddb1f242b3"} Dec 10 14:42:31 crc kubenswrapper[4847]: I1210 14:42:31.011154 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:42:31 crc kubenswrapper[4847]: I1210 14:42:31.011793 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.015690 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czhlz" event={"ID":"32400504-9741-4abf-bdbf-a8ea9fcb8e4e","Type":"ContainerStarted","Data":"cbbbc02925db69af15abd144ed9d350cb7ada6712454fc52a9a48c0cb9f476ad"} Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.020050 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"c8ca3d480d943eef0fe4ca8a2c52ea26a2f0be2796caada98af272fb2eadb0c1"} Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.020111 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"9e870ebd1e01a16e5b5ecf58666b50f859404fc43f7ba7537b6cf1c2733f1d9f"} Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.020125 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"184d4e6a3f618b43f57c459f7f37dc9122949c3aa95e171bd46bfc36376f2352"} Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.020137 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"7f529edf15d4f3a694e1693e8657964f29bf654dcc598055a6f8354ddda78235"} Dec 10 14:42:34 crc kubenswrapper[4847]: I1210 14:42:34.032925 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-czhlz" podStartSLOduration=2.277441519 podStartE2EDuration="33.032906213s" podCreationTimestamp="2025-12-10 14:42:01 +0000 UTC" firstStartedPulling="2025-12-10 14:42:02.225554097 +0000 UTC m=+1071.794771727" lastFinishedPulling="2025-12-10 14:42:32.981018791 +0000 UTC m=+1102.550236421" observedRunningTime="2025-12-10 14:42:34.031495903 +0000 UTC m=+1103.600713543" watchObservedRunningTime="2025-12-10 14:42:34.032906213 +0000 UTC m=+1103.602123853" Dec 10 14:42:35 crc kubenswrapper[4847]: I1210 14:42:35.030129 4847 generic.go:334] "Generic (PLEG): container finished" podID="ecf994d1-33cd-404a-bec2-01956b20b78d" containerID="8347dbf81b888a31f98854a1ff7e74c2100b2619c2a2325b51d66e1270c79abf" exitCode=0 Dec 10 14:42:35 crc kubenswrapper[4847]: I1210 14:42:35.030201 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ssqhj" event={"ID":"ecf994d1-33cd-404a-bec2-01956b20b78d","Type":"ContainerDied","Data":"8347dbf81b888a31f98854a1ff7e74c2100b2619c2a2325b51d66e1270c79abf"} Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.402646 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.560827 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle\") pod \"ecf994d1-33cd-404a-bec2-01956b20b78d\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.560956 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data\") pod \"ecf994d1-33cd-404a-bec2-01956b20b78d\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.561022 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbjz4\" (UniqueName: \"kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4\") pod \"ecf994d1-33cd-404a-bec2-01956b20b78d\" (UID: \"ecf994d1-33cd-404a-bec2-01956b20b78d\") " Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.569039 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4" (OuterVolumeSpecName: "kube-api-access-dbjz4") pod "ecf994d1-33cd-404a-bec2-01956b20b78d" (UID: "ecf994d1-33cd-404a-bec2-01956b20b78d"). InnerVolumeSpecName "kube-api-access-dbjz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.586559 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecf994d1-33cd-404a-bec2-01956b20b78d" (UID: "ecf994d1-33cd-404a-bec2-01956b20b78d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.606672 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data" (OuterVolumeSpecName: "config-data") pod "ecf994d1-33cd-404a-bec2-01956b20b78d" (UID: "ecf994d1-33cd-404a-bec2-01956b20b78d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.673230 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.673302 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbjz4\" (UniqueName: \"kubernetes.io/projected/ecf994d1-33cd-404a-bec2-01956b20b78d-kube-api-access-dbjz4\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:36 crc kubenswrapper[4847]: I1210 14:42:36.673316 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf994d1-33cd-404a-bec2-01956b20b78d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.046820 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ssqhj" event={"ID":"ecf994d1-33cd-404a-bec2-01956b20b78d","Type":"ContainerDied","Data":"fb9ac85c1cffaa567101621a05ce9353dc06fd2ce925a7fc676e25c5589a60ef"} Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.047157 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb9ac85c1cffaa567101621a05ce9353dc06fd2ce925a7fc676e25c5589a60ef" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.046878 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ssqhj" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.347692 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348114 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b75e08-4a3d-410f-9c10-81f695250989" containerName="ovn-config" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348132 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b75e08-4a3d-410f-9c10-81f695250989" containerName="ovn-config" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348148 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bbab02-b41c-4b58-a235-0c88b7eeb1b3" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348157 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bbab02-b41c-4b58-a235-0c88b7eeb1b3" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348169 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11592b8-0374-4b1d-a579-64b72830148a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348177 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11592b8-0374-4b1d-a579-64b72830148a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348193 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7148ddb7-2803-41ab-bf95-f820c60614d1" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348200 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7148ddb7-2803-41ab-bf95-f820c60614d1" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348211 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bba7667-cbb4-4636-9197-2409e86b8cde" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348218 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bba7667-cbb4-4636-9197-2409e86b8cde" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348241 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5096792-9e2e-4843-85da-9e14112d056a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348249 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5096792-9e2e-4843-85da-9e14112d056a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348263 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b75aa68-b977-4614-877b-9828d3045226" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348270 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b75aa68-b977-4614-877b-9828d3045226" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: E1210 14:42:37.348284 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecf994d1-33cd-404a-bec2-01956b20b78d" containerName="keystone-db-sync" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348291 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecf994d1-33cd-404a-bec2-01956b20b78d" containerName="keystone-db-sync" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348516 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b75aa68-b977-4614-877b-9828d3045226" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348539 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bba7667-cbb4-4636-9197-2409e86b8cde" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348568 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7148ddb7-2803-41ab-bf95-f820c60614d1" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348578 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="17bbab02-b41c-4b58-a235-0c88b7eeb1b3" containerName="mariadb-database-create" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348593 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5096792-9e2e-4843-85da-9e14112d056a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348646 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b75e08-4a3d-410f-9c10-81f695250989" containerName="ovn-config" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348662 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11592b8-0374-4b1d-a579-64b72830148a" containerName="mariadb-account-create-update" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.348676 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecf994d1-33cd-404a-bec2-01956b20b78d" containerName="keystone-db-sync" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.350208 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.385903 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.386981 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.387111 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.387210 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.387369 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdsv2\" (UniqueName: \"kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.387450 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.395279 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ltt7z"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.396661 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.399144 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.399381 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mz28n" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.399674 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.400252 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.400500 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.421790 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ltt7z"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.491651 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.491730 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495008 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495076 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47tvc\" (UniqueName: \"kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495120 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495178 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495197 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495239 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495274 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495293 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdsv2\" (UniqueName: \"kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.495334 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.506501 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.506508 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.507989 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.523319 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-twlv7"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.524489 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.532246 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.534475 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.534793 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.536280 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-blk76" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.575877 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdsv2\" (UniqueName: \"kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2\") pod \"dnsmasq-dns-5c9d85d47c-29lw7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.579339 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-twlv7"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.598900 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.598969 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599008 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599041 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599096 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599142 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599169 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599198 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599241 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47tvc\" (UniqueName: \"kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599298 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.599351 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg8k4\" (UniqueName: \"kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.607200 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.611385 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.613910 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.613969 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.624937 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.625151 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.632066 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.632181 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.632303 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-7zzgs" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.632373 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.637307 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.647689 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47tvc\" (UniqueName: \"kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc\") pod \"keystone-bootstrap-ltt7z\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.668152 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704612 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704666 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704701 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704761 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs79n\" (UniqueName: \"kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704777 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704812 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704833 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg8k4\" (UniqueName: \"kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704850 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704877 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704898 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704920 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.704993 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.708458 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.715021 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.715458 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.720262 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.726204 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.745432 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg8k4\" (UniqueName: \"kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.745889 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts\") pod \"cinder-db-sync-twlv7\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.783782 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-66gqz"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.785081 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.792769 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.793017 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dfjp4" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.793141 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.811905 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs79n\" (UniqueName: \"kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.811995 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812031 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812063 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812106 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812133 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812210 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.812250 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g2fc\" (UniqueName: \"kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.817655 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.826114 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.832322 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.840816 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.869944 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs79n\" (UniqueName: \"kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n\") pod \"horizon-7985b45c4c-mt5vg\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.881396 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-twlv7" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.904976 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-66gqz"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.905301 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.913437 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.913523 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.913629 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g2fc\" (UniqueName: \"kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.920646 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.921616 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.939036 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g2fc\" (UniqueName: \"kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc\") pod \"neutron-db-sync-66gqz\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.943770 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:37 crc kubenswrapper[4847]: I1210 14:42:37.965094 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.002264 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.004449 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.067705 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.113365 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.113473 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.113570 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.113632 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t89nl\" (UniqueName: \"kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.113972 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.114570 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.118677 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.118951 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.121675 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.122785 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.134647 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-j5hjt"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.140298 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.146050 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.146229 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nf2pv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.146380 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.156898 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.173558 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-p6r7l"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.174967 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.178009 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rlqtm" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.178255 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.184446 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j5hjt"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.194690 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.212990 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p6r7l"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215401 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215458 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215499 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215531 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215552 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhvh6\" (UniqueName: \"kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215572 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215595 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmx4p\" (UniqueName: \"kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215621 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215656 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215686 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215724 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215753 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215776 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215806 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215833 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpcb4\" (UniqueName: \"kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215861 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t89nl\" (UniqueName: \"kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215896 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215918 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215948 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.215979 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.216030 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgxln\" (UniqueName: \"kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.216051 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.216071 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.216097 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.216156 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.217537 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.218276 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.220290 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.220993 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.234205 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-66gqz" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.278480 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t89nl\" (UniqueName: \"kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl\") pod \"dnsmasq-dns-6ffb94d8ff-n7j7t\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318058 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgxln\" (UniqueName: \"kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318099 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318121 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318137 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318165 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318209 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318238 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318257 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318276 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhvh6\" (UniqueName: \"kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318295 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318321 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmx4p\" (UniqueName: \"kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318373 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318389 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318411 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318435 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318467 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpcb4\" (UniqueName: \"kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318504 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318555 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.318574 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.326429 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.326521 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.327094 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.327224 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.328096 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.328488 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.328603 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.328736 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.333348 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.341110 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.341132 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.341872 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.342169 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.343203 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.343216 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.346636 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmx4p\" (UniqueName: \"kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p\") pod \"barbican-db-sync-p6r7l\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.346767 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgxln\" (UniqueName: \"kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln\") pod \"ceilometer-0\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.349364 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpcb4\" (UniqueName: \"kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4\") pod \"horizon-7f696c998f-hsmnv\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.355335 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhvh6\" (UniqueName: \"kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.370824 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data\") pod \"placement-db-sync-j5hjt\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.506135 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.536640 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.569640 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.577066 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ltt7z"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.599160 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j5hjt" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.627273 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.878006 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:38 crc kubenswrapper[4847]: I1210 14:42:38.938323 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-twlv7"] Dec 10 14:42:39 crc kubenswrapper[4847]: W1210 14:42:39.048020 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82ca7b48_ad83_46b4_a813_de1ec043a3e2.slice/crio-97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176 WatchSource:0}: Error finding container 97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176: Status 404 returned error can't find the container with id 97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176 Dec 10 14:42:39 crc kubenswrapper[4847]: W1210 14:42:39.051841 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6995af57_136d_4e9a_945f_ce4d3dbdb5d7.slice/crio-748cc7a0d7ddb08e423fbf1394bc05a33c07d34fac09633103607e4ae9eb5b7c WatchSource:0}: Error finding container 748cc7a0d7ddb08e423fbf1394bc05a33c07d34fac09633103607e4ae9eb5b7c: Status 404 returned error can't find the container with id 748cc7a0d7ddb08e423fbf1394bc05a33c07d34fac09633103607e4ae9eb5b7c Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.078467 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:42:39 crc kubenswrapper[4847]: W1210 14:42:39.105046 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8900755_ba7e_4e00_9500_69afbdb69064.slice/crio-6666dd22527733969080f6528ad6fe1ff475054cde5c0e9ab2dfae460bf49e9d WatchSource:0}: Error finding container 6666dd22527733969080f6528ad6fe1ff475054cde5c0e9ab2dfae460bf49e9d: Status 404 returned error can't find the container with id 6666dd22527733969080f6528ad6fe1ff475054cde5c0e9ab2dfae460bf49e9d Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.184491 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" event={"ID":"6995af57-136d-4e9a-945f-ce4d3dbdb5d7","Type":"ContainerStarted","Data":"748cc7a0d7ddb08e423fbf1394bc05a33c07d34fac09633103607e4ae9eb5b7c"} Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.199527 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7985b45c4c-mt5vg" event={"ID":"d8900755-ba7e-4e00-9500-69afbdb69064","Type":"ContainerStarted","Data":"6666dd22527733969080f6528ad6fe1ff475054cde5c0e9ab2dfae460bf49e9d"} Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.201940 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-66gqz"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.203421 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-twlv7" event={"ID":"82ca7b48-ad83-46b4-a813-de1ec043a3e2","Type":"ContainerStarted","Data":"97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176"} Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.205047 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ltt7z" event={"ID":"a3eaa9cb-2587-47ea-b174-4bbf895acfe1","Type":"ContainerStarted","Data":"49a926c5567683be389028a32df04f7ef25c2fc2b667f27d9566203f5ce5d4b6"} Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.464056 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.644742 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.670904 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-j5hjt"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.681524 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.871788 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.880562 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p6r7l"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.948733 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:42:39 crc kubenswrapper[4847]: I1210 14:42:39.998682 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.001149 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.006363 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.091911 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.092043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.092280 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.092379 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.092534 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4jx5\" (UniqueName: \"kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.193953 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.194000 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.194048 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4jx5\" (UniqueName: \"kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.194071 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.194107 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.194852 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.195447 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.195463 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.200458 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.214821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ltt7z" event={"ID":"a3eaa9cb-2587-47ea-b174-4bbf895acfe1","Type":"ContainerStarted","Data":"7399057319f9de26ea7513b978760aa2f589a8804d8990d90a7990641b0fa320"} Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.219741 4847 generic.go:334] "Generic (PLEG): container finished" podID="6995af57-136d-4e9a-945f-ce4d3dbdb5d7" containerID="ae65fa6f3e3bb48ace504792fd29439ee717f73e2a1908bc75833a17814be86e" exitCode=0 Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.219812 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" event={"ID":"6995af57-136d-4e9a-945f-ce4d3dbdb5d7","Type":"ContainerDied","Data":"ae65fa6f3e3bb48ace504792fd29439ee717f73e2a1908bc75833a17814be86e"} Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.220077 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4jx5\" (UniqueName: \"kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5\") pod \"horizon-5f98d7b94c-594v6\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.224376 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-66gqz" event={"ID":"ea4071c9-d190-4706-8f11-16b77f3db069","Type":"ContainerStarted","Data":"6ec3a10674aa81d3e1e8db5c312a2b5019e8ebd2983178875a1545c9454cdffa"} Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.224427 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-66gqz" event={"ID":"ea4071c9-d190-4706-8f11-16b77f3db069","Type":"ContainerStarted","Data":"50a93a0db345f9a038b26782d2e9368b214186e8abb90abad093db70bb0a39d4"} Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.246556 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ltt7z" podStartSLOduration=3.246526174 podStartE2EDuration="3.246526174s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:40.239638993 +0000 UTC m=+1109.808856623" watchObservedRunningTime="2025-12-10 14:42:40.246526174 +0000 UTC m=+1109.815743804" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.286469 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-66gqz" podStartSLOduration=3.286440465 podStartE2EDuration="3.286440465s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:40.284247294 +0000 UTC m=+1109.853464924" watchObservedRunningTime="2025-12-10 14:42:40.286440465 +0000 UTC m=+1109.855658095" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.327285 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:42:40 crc kubenswrapper[4847]: W1210 14:42:40.375013 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ce653b9_194a_4390_9a40_241ff10a18df.slice/crio-06bf37941318a4e7e224a5c675f29b21f16e8a0bf2f1f85a3a3d815f45b1799c WatchSource:0}: Error finding container 06bf37941318a4e7e224a5c675f29b21f16e8a0bf2f1f85a3a3d815f45b1799c: Status 404 returned error can't find the container with id 06bf37941318a4e7e224a5c675f29b21f16e8a0bf2f1f85a3a3d815f45b1799c Dec 10 14:42:40 crc kubenswrapper[4847]: W1210 14:42:40.377588 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e3440ea_24b3_4cc9_9872_8c4e28ab340b.slice/crio-505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2 WatchSource:0}: Error finding container 505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2: Status 404 returned error can't find the container with id 505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2 Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.756408 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.808146 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config\") pod \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.808245 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc\") pod \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.808284 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb\") pod \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.808320 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdsv2\" (UniqueName: \"kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2\") pod \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.808414 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb\") pod \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\" (UID: \"6995af57-136d-4e9a-945f-ce4d3dbdb5d7\") " Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.840584 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6995af57-136d-4e9a-945f-ce4d3dbdb5d7" (UID: "6995af57-136d-4e9a-945f-ce4d3dbdb5d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.844755 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2" (OuterVolumeSpecName: "kube-api-access-jdsv2") pod "6995af57-136d-4e9a-945f-ce4d3dbdb5d7" (UID: "6995af57-136d-4e9a-945f-ce4d3dbdb5d7"). InnerVolumeSpecName "kube-api-access-jdsv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.876450 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6995af57-136d-4e9a-945f-ce4d3dbdb5d7" (UID: "6995af57-136d-4e9a-945f-ce4d3dbdb5d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.895470 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config" (OuterVolumeSpecName: "config") pod "6995af57-136d-4e9a-945f-ce4d3dbdb5d7" (UID: "6995af57-136d-4e9a-945f-ce4d3dbdb5d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.930187 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.930234 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.930246 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.930258 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdsv2\" (UniqueName: \"kubernetes.io/projected/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-kube-api-access-jdsv2\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:40 crc kubenswrapper[4847]: I1210 14:42:40.934446 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6995af57-136d-4e9a-945f-ce4d3dbdb5d7" (UID: "6995af57-136d-4e9a-945f-ce4d3dbdb5d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.031750 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6995af57-136d-4e9a-945f-ce4d3dbdb5d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.162546 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.259045 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f696c998f-hsmnv" event={"ID":"c58022e7-62a2-4ef5-bce1-024d8da0b9a8","Type":"ContainerStarted","Data":"53dd00abfff2200847827b0c57311e05154de6f2c82e248f2d2c793475386a45"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.262971 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p6r7l" event={"ID":"66a6b89e-e44e-4e19-92bd-12d58ad99fd1","Type":"ContainerStarted","Data":"ed765f016f60dad6ce8022f4ba48e2f04dc240731f80f8bbe1382d00f7b028ef"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.270286 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f98d7b94c-594v6" event={"ID":"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7","Type":"ContainerStarted","Data":"ed736f92ebe1cad9d5bbe2c8f8de9de2291122113684f095d39d329432711448"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.279801 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j5hjt" event={"ID":"8e3440ea-24b3-4cc9-9872-8c4e28ab340b","Type":"ContainerStarted","Data":"505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.287892 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerStarted","Data":"169558239df569f96e448d76dd494e20d5f05d2e86385509862f4f534f53b4cb"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.294515 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" event={"ID":"2ce653b9-194a-4390-9a40-241ff10a18df","Type":"ContainerStarted","Data":"06bf37941318a4e7e224a5c675f29b21f16e8a0bf2f1f85a3a3d815f45b1799c"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.299168 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.305980 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-29lw7" event={"ID":"6995af57-136d-4e9a-945f-ce4d3dbdb5d7","Type":"ContainerDied","Data":"748cc7a0d7ddb08e423fbf1394bc05a33c07d34fac09633103607e4ae9eb5b7c"} Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.306073 4847 scope.go:117] "RemoveContainer" containerID="ae65fa6f3e3bb48ace504792fd29439ee717f73e2a1908bc75833a17814be86e" Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.410970 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:41 crc kubenswrapper[4847]: I1210 14:42:41.437911 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-29lw7"] Dec 10 14:42:42 crc kubenswrapper[4847]: I1210 14:42:42.349685 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"7ba210fcd4a48a72e737cf3d384c57576d536b601b98da0f52f58e2e1141a55a"} Dec 10 14:42:42 crc kubenswrapper[4847]: I1210 14:42:42.350045 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"d83e5dedbf9ddc822ed306e0706031b64a26d0ff0eb81788d2b2ac8b03edfd1f"} Dec 10 14:42:42 crc kubenswrapper[4847]: I1210 14:42:42.367953 4847 generic.go:334] "Generic (PLEG): container finished" podID="2ce653b9-194a-4390-9a40-241ff10a18df" containerID="d1849a7699352f6e2848c06dd0455ebc6b00e31bbe5af200d88b6de6faecba81" exitCode=0 Dec 10 14:42:42 crc kubenswrapper[4847]: I1210 14:42:42.368085 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" event={"ID":"2ce653b9-194a-4390-9a40-241ff10a18df","Type":"ContainerDied","Data":"d1849a7699352f6e2848c06dd0455ebc6b00e31bbe5af200d88b6de6faecba81"} Dec 10 14:42:42 crc kubenswrapper[4847]: I1210 14:42:42.773131 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6995af57-136d-4e9a-945f-ce4d3dbdb5d7" path="/var/lib/kubelet/pods/6995af57-136d-4e9a-945f-ce4d3dbdb5d7/volumes" Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.466830 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" event={"ID":"2ce653b9-194a-4390-9a40-241ff10a18df","Type":"ContainerStarted","Data":"186cd89ae18ada42feeaa7786419b47e5b7c320f731897476caded86505873e4"} Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.468827 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.487096 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"fd09f940c242e2f8f99913a1f4784daf5326ae1708708bf053b6b04e3c13011e"} Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.487144 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"df2d4b8447b46515f0724df72386d592c1f19e515e9e50c3b480e0769f83a294"} Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.487156 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"23c2d8366647302e7e65edd4f5710d9f1b99c7bc350ff981f9348d0400cf3730"} Dec 10 14:42:43 crc kubenswrapper[4847]: I1210 14:42:43.497474 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" podStartSLOduration=6.497446106 podStartE2EDuration="6.497446106s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:42:43.492189771 +0000 UTC m=+1113.061407401" watchObservedRunningTime="2025-12-10 14:42:43.497446106 +0000 UTC m=+1113.066663736" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.513512 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"4692c4535cee66953c82e6ad562360772d7a0fc8ff54fe179097592e8d0edc30"} Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.513906 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"88c96cdf-8002-4829-9f8c-3abfc8315722","Type":"ContainerStarted","Data":"a0253bf8c1f551e8c01208254a18a148d7acece1fd7ffdae25a782f738c5fc2d"} Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.548328 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=46.24015461 podStartE2EDuration="1m7.548305208s" podCreationTimestamp="2025-12-10 14:41:37 +0000 UTC" firstStartedPulling="2025-12-10 14:42:19.198887846 +0000 UTC m=+1088.768105476" lastFinishedPulling="2025-12-10 14:42:40.507038444 +0000 UTC m=+1110.076256074" observedRunningTime="2025-12-10 14:42:44.543574249 +0000 UTC m=+1114.112791899" watchObservedRunningTime="2025-12-10 14:42:44.548305208 +0000 UTC m=+1114.117522838" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.820215 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.845149 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:42:44 crc kubenswrapper[4847]: E1210 14:42:44.845523 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6995af57-136d-4e9a-945f-ce4d3dbdb5d7" containerName="init" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.845539 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6995af57-136d-4e9a-945f-ce4d3dbdb5d7" containerName="init" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.845823 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6995af57-136d-4e9a-945f-ce4d3dbdb5d7" containerName="init" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.846659 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.849140 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.881779 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.937746 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-442qv\" (UniqueName: \"kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.937799 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.937820 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.937846 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.937865 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:44 crc kubenswrapper[4847]: I1210 14:42:44.938118 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.049178 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.053952 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-442qv\" (UniqueName: \"kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.054085 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.054125 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.054205 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.054248 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.055523 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.056233 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.057249 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.057949 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.058593 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.079840 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-442qv\" (UniqueName: \"kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv\") pod \"dnsmasq-dns-cf78879c9-d7xkj\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.164491 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.526279 4847 generic.go:334] "Generic (PLEG): container finished" podID="a3eaa9cb-2587-47ea-b174-4bbf895acfe1" containerID="7399057319f9de26ea7513b978760aa2f589a8804d8990d90a7990641b0fa320" exitCode=0 Dec 10 14:42:45 crc kubenswrapper[4847]: I1210 14:42:45.526317 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ltt7z" event={"ID":"a3eaa9cb-2587-47ea-b174-4bbf895acfe1","Type":"ContainerDied","Data":"7399057319f9de26ea7513b978760aa2f589a8804d8990d90a7990641b0fa320"} Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.142612 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.193487 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.195180 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.200270 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.216360 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.265982 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.278923 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279156 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279274 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279332 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7n5r\" (UniqueName: \"kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279422 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279506 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.279602 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.309163 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d4f6b7576-wvhs4"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.311387 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.325342 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f6b7576-wvhs4"] Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.382943 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7n5r\" (UniqueName: \"kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383008 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-scripts\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383045 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-combined-ca-bundle\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383190 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383226 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxkgz\" (UniqueName: \"kubernetes.io/projected/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-kube-api-access-xxkgz\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383258 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-secret-key\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383321 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383400 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-logs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383429 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383462 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-tls-certs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383726 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383810 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-config-data\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383860 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.383947 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.384322 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.384464 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.385750 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.392265 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.396012 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.402561 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7n5r\" (UniqueName: \"kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.408650 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs\") pod \"horizon-8676865fc8-rvrhh\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484564 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-config-data\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484632 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-scripts\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484657 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-combined-ca-bundle\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484688 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxkgz\" (UniqueName: \"kubernetes.io/projected/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-kube-api-access-xxkgz\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484726 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-secret-key\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484762 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-logs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.484789 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-tls-certs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.485787 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-logs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.486133 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-scripts\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.486502 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-config-data\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.491121 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-combined-ca-bundle\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.491132 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-secret-key\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.491210 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-horizon-tls-certs\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.504703 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxkgz\" (UniqueName: \"kubernetes.io/projected/f7e7178d-cd26-4623-a05c-6eee95cb5dd6-kube-api-access-xxkgz\") pod \"horizon-6d4f6b7576-wvhs4\" (UID: \"f7e7178d-cd26-4623-a05c-6eee95cb5dd6\") " pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.531119 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.536792 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" containerID="cri-o://186cd89ae18ada42feeaa7786419b47e5b7c320f731897476caded86505873e4" gracePeriod=10 Dec 10 14:42:46 crc kubenswrapper[4847]: I1210 14:42:46.646628 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:42:47 crc kubenswrapper[4847]: I1210 14:42:47.558245 4847 generic.go:334] "Generic (PLEG): container finished" podID="2ce653b9-194a-4390-9a40-241ff10a18df" containerID="186cd89ae18ada42feeaa7786419b47e5b7c320f731897476caded86505873e4" exitCode=0 Dec 10 14:42:47 crc kubenswrapper[4847]: I1210 14:42:47.558341 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" event={"ID":"2ce653b9-194a-4390-9a40-241ff10a18df","Type":"ContainerDied","Data":"186cd89ae18ada42feeaa7786419b47e5b7c320f731897476caded86505873e4"} Dec 10 14:42:48 crc kubenswrapper[4847]: I1210 14:42:48.509554 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Dec 10 14:42:48 crc kubenswrapper[4847]: I1210 14:42:48.933439 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033293 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033433 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47tvc\" (UniqueName: \"kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033545 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033571 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033618 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.033634 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts\") pod \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\" (UID: \"a3eaa9cb-2587-47ea-b174-4bbf895acfe1\") " Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.045241 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.045512 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc" (OuterVolumeSpecName: "kube-api-access-47tvc") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "kube-api-access-47tvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.049951 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts" (OuterVolumeSpecName: "scripts") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.052976 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.065840 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.069129 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data" (OuterVolumeSpecName: "config-data") pod "a3eaa9cb-2587-47ea-b174-4bbf895acfe1" (UID: "a3eaa9cb-2587-47ea-b174-4bbf895acfe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136075 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136110 4847 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136120 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136129 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136141 4847 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.136152 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47tvc\" (UniqueName: \"kubernetes.io/projected/a3eaa9cb-2587-47ea-b174-4bbf895acfe1-kube-api-access-47tvc\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.578885 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ltt7z" event={"ID":"a3eaa9cb-2587-47ea-b174-4bbf895acfe1","Type":"ContainerDied","Data":"49a926c5567683be389028a32df04f7ef25c2fc2b667f27d9566203f5ce5d4b6"} Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.578943 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49a926c5567683be389028a32df04f7ef25c2fc2b667f27d9566203f5ce5d4b6" Dec 10 14:42:49 crc kubenswrapper[4847]: I1210 14:42:49.578960 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ltt7z" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.035815 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ltt7z"] Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.048801 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ltt7z"] Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.137373 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fw2fp"] Dec 10 14:42:50 crc kubenswrapper[4847]: E1210 14:42:50.137804 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3eaa9cb-2587-47ea-b174-4bbf895acfe1" containerName="keystone-bootstrap" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.137820 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3eaa9cb-2587-47ea-b174-4bbf895acfe1" containerName="keystone-bootstrap" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.138044 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3eaa9cb-2587-47ea-b174-4bbf895acfe1" containerName="keystone-bootstrap" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.138594 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.140429 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.142524 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.142696 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.143427 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mz28n" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.147186 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.171799 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fw2fp"] Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258482 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258532 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258557 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25x68\" (UniqueName: \"kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258631 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258657 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.258692 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.360580 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.360940 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.361015 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.361115 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.361137 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.361163 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25x68\" (UniqueName: \"kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.367865 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.369273 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.376658 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.378765 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.385322 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.396953 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25x68\" (UniqueName: \"kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68\") pod \"keystone-bootstrap-fw2fp\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.470141 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.595077 4847 generic.go:334] "Generic (PLEG): container finished" podID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" containerID="cbbbc02925db69af15abd144ed9d350cb7ada6712454fc52a9a48c0cb9f476ad" exitCode=0 Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.595218 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czhlz" event={"ID":"32400504-9741-4abf-bdbf-a8ea9fcb8e4e","Type":"ContainerDied","Data":"cbbbc02925db69af15abd144ed9d350cb7ada6712454fc52a9a48c0cb9f476ad"} Dec 10 14:42:50 crc kubenswrapper[4847]: I1210 14:42:50.790876 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3eaa9cb-2587-47ea-b174-4bbf895acfe1" path="/var/lib/kubelet/pods/a3eaa9cb-2587-47ea-b174-4bbf895acfe1/volumes" Dec 10 14:42:58 crc kubenswrapper[4847]: I1210 14:42:58.507413 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.293005 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.293203 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dfh67chc4h95h557h5fdh5f7h668h55fh85h6ch597h669h684h5d8h55bh5cch7ch675hcdh86h54hcbh576h78h565h68fh548h578h676h5b4hdfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cpcb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f696c998f-hsmnv_openstack(c58022e7-62a2-4ef5-bce1-024d8da0b9a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.298443 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f696c998f-hsmnv" podUID="c58022e7-62a2-4ef5-bce1-024d8da0b9a8" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.316090 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.316300 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d4hdfh558h57ch69h5d9hfdh577h7ch65bh9dh85h55ch6dh56dh545h549h599h59ch69h694h5b5h5bh5f8h647h577h5dch5bh64bh74h66dh58bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bs79n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7985b45c4c-mt5vg_openstack(d8900755-ba7e-4e00-9500-69afbdb69064): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.322733 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7985b45c4c-mt5vg" podUID="d8900755-ba7e-4e00-9500-69afbdb69064" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.325680 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.325891 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd8h546h97h9dhfch595h6ch55dhdfh5c5h569h5dch658h5cdh696h5bbh686h5d5h5cbhddhfdh5d8h55dhf9h54dh59bh66ch8hc6h78h577h594q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x4jx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5f98d7b94c-594v6_openstack(fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:42:59 crc kubenswrapper[4847]: E1210 14:42:59.328702 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5f98d7b94c-594v6" podUID="fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.429366 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.454214 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb\") pod \"2ce653b9-194a-4390-9a40-241ff10a18df\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.454287 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc\") pod \"2ce653b9-194a-4390-9a40-241ff10a18df\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.454362 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb\") pod \"2ce653b9-194a-4390-9a40-241ff10a18df\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.454439 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config\") pod \"2ce653b9-194a-4390-9a40-241ff10a18df\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.454508 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t89nl\" (UniqueName: \"kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl\") pod \"2ce653b9-194a-4390-9a40-241ff10a18df\" (UID: \"2ce653b9-194a-4390-9a40-241ff10a18df\") " Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.461536 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl" (OuterVolumeSpecName: "kube-api-access-t89nl") pod "2ce653b9-194a-4390-9a40-241ff10a18df" (UID: "2ce653b9-194a-4390-9a40-241ff10a18df"). InnerVolumeSpecName "kube-api-access-t89nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.507463 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ce653b9-194a-4390-9a40-241ff10a18df" (UID: "2ce653b9-194a-4390-9a40-241ff10a18df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.510810 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config" (OuterVolumeSpecName: "config") pod "2ce653b9-194a-4390-9a40-241ff10a18df" (UID: "2ce653b9-194a-4390-9a40-241ff10a18df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.512979 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ce653b9-194a-4390-9a40-241ff10a18df" (UID: "2ce653b9-194a-4390-9a40-241ff10a18df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.516033 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ce653b9-194a-4390-9a40-241ff10a18df" (UID: "2ce653b9-194a-4390-9a40-241ff10a18df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.556516 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t89nl\" (UniqueName: \"kubernetes.io/projected/2ce653b9-194a-4390-9a40-241ff10a18df-kube-api-access-t89nl\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.556553 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.556564 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.556573 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.556581 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ce653b9-194a-4390-9a40-241ff10a18df-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.673522 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" event={"ID":"2ce653b9-194a-4390-9a40-241ff10a18df","Type":"ContainerDied","Data":"06bf37941318a4e7e224a5c675f29b21f16e8a0bf2f1f85a3a3d815f45b1799c"} Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.673580 4847 scope.go:117] "RemoveContainer" containerID="186cd89ae18ada42feeaa7786419b47e5b7c320f731897476caded86505873e4" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.673692 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.738342 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:42:59 crc kubenswrapper[4847]: I1210 14:42:59.750296 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-n7j7t"] Dec 10 14:43:00 crc kubenswrapper[4847]: I1210 14:43:00.774142 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" path="/var/lib/kubelet/pods/2ce653b9-194a-4390-9a40-241ff10a18df/volumes" Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.011459 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.011534 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.011605 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.012539 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.012617 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7" gracePeriod=600 Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.691900 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7" exitCode=0 Dec 10 14:43:01 crc kubenswrapper[4847]: I1210 14:43:01.691945 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7"} Dec 10 14:43:03 crc kubenswrapper[4847]: I1210 14:43:03.510682 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6ffb94d8ff-n7j7t" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 10 14:43:03 crc kubenswrapper[4847]: I1210 14:43:03.713092 4847 generic.go:334] "Generic (PLEG): container finished" podID="ea4071c9-d190-4706-8f11-16b77f3db069" containerID="6ec3a10674aa81d3e1e8db5c312a2b5019e8ebd2983178875a1545c9454cdffa" exitCode=0 Dec 10 14:43:03 crc kubenswrapper[4847]: I1210 14:43:03.713136 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-66gqz" event={"ID":"ea4071c9-d190-4706-8f11-16b77f3db069","Type":"ContainerDied","Data":"6ec3a10674aa81d3e1e8db5c312a2b5019e8ebd2983178875a1545c9454cdffa"} Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.516575 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czhlz" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.565162 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data\") pod \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.565227 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle\") pod \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.565289 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data\") pod \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.565320 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spcht\" (UniqueName: \"kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht\") pod \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\" (UID: \"32400504-9741-4abf-bdbf-a8ea9fcb8e4e\") " Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.576636 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32400504-9741-4abf-bdbf-a8ea9fcb8e4e" (UID: "32400504-9741-4abf-bdbf-a8ea9fcb8e4e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.578169 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht" (OuterVolumeSpecName: "kube-api-access-spcht") pod "32400504-9741-4abf-bdbf-a8ea9fcb8e4e" (UID: "32400504-9741-4abf-bdbf-a8ea9fcb8e4e"). InnerVolumeSpecName "kube-api-access-spcht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.611469 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32400504-9741-4abf-bdbf-a8ea9fcb8e4e" (UID: "32400504-9741-4abf-bdbf-a8ea9fcb8e4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.625490 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data" (OuterVolumeSpecName: "config-data") pod "32400504-9741-4abf-bdbf-a8ea9fcb8e4e" (UID: "32400504-9741-4abf-bdbf-a8ea9fcb8e4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.667988 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.668032 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spcht\" (UniqueName: \"kubernetes.io/projected/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-kube-api-access-spcht\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.668047 4847 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.668058 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32400504-9741-4abf-bdbf-a8ea9fcb8e4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.769865 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-czhlz" Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.770298 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-czhlz" event={"ID":"32400504-9741-4abf-bdbf-a8ea9fcb8e4e","Type":"ContainerDied","Data":"7bb8148a26e0c12e9fe13cd8d060b6abc265a52bdaa0fd9fdff50a16f9d81218"} Dec 10 14:43:10 crc kubenswrapper[4847]: I1210 14:43:10.770337 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bb8148a26e0c12e9fe13cd8d060b6abc265a52bdaa0fd9fdff50a16f9d81218" Dec 10 14:43:11 crc kubenswrapper[4847]: E1210 14:43:11.135142 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 10 14:43:11 crc kubenswrapper[4847]: E1210 14:43:11.135308 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfbhb6h98h67fh678h58bh556h5bfh84h695h5d9h8chd4hbbh5b6hbch5c5h596h565h5c8h57h5fh5bh5f9h679h88h66dh557hfch5c4h595h648q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgxln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(b9ecc8fb-2094-44fd-b157-c654a293036e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.175099 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.183839 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.204009 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.215628 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-66gqz" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291301 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs\") pod \"d8900755-ba7e-4e00-9500-69afbdb69064\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291614 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data\") pod \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291670 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4jx5\" (UniqueName: \"kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5\") pod \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291740 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs\") pod \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291800 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key\") pod \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291827 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs79n\" (UniqueName: \"kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n\") pod \"d8900755-ba7e-4e00-9500-69afbdb69064\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291853 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key\") pod \"d8900755-ba7e-4e00-9500-69afbdb69064\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291888 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpcb4\" (UniqueName: \"kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4\") pod \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291916 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config\") pod \"ea4071c9-d190-4706-8f11-16b77f3db069\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291953 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs\") pod \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.291978 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data\") pod \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292210 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts\") pod \"d8900755-ba7e-4e00-9500-69afbdb69064\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292265 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key\") pod \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292294 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data" (OuterVolumeSpecName: "config-data") pod "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" (UID: "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292331 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts\") pod \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\" (UID: \"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292373 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle\") pod \"ea4071c9-d190-4706-8f11-16b77f3db069\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292460 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts\") pod \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\" (UID: \"c58022e7-62a2-4ef5-bce1-024d8da0b9a8\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292486 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g2fc\" (UniqueName: \"kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc\") pod \"ea4071c9-d190-4706-8f11-16b77f3db069\" (UID: \"ea4071c9-d190-4706-8f11-16b77f3db069\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.292525 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data\") pod \"d8900755-ba7e-4e00-9500-69afbdb69064\" (UID: \"d8900755-ba7e-4e00-9500-69afbdb69064\") " Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.293082 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs" (OuterVolumeSpecName: "logs") pod "d8900755-ba7e-4e00-9500-69afbdb69064" (UID: "d8900755-ba7e-4e00-9500-69afbdb69064"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.293601 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data" (OuterVolumeSpecName: "config-data") pod "c58022e7-62a2-4ef5-bce1-024d8da0b9a8" (UID: "c58022e7-62a2-4ef5-bce1-024d8da0b9a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.293781 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs" (OuterVolumeSpecName: "logs") pod "c58022e7-62a2-4ef5-bce1-024d8da0b9a8" (UID: "c58022e7-62a2-4ef5-bce1-024d8da0b9a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.293790 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs" (OuterVolumeSpecName: "logs") pod "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" (UID: "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.293971 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts" (OuterVolumeSpecName: "scripts") pod "d8900755-ba7e-4e00-9500-69afbdb69064" (UID: "d8900755-ba7e-4e00-9500-69afbdb69064"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.294102 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts" (OuterVolumeSpecName: "scripts") pod "c58022e7-62a2-4ef5-bce1-024d8da0b9a8" (UID: "c58022e7-62a2-4ef5-bce1-024d8da0b9a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.294375 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data" (OuterVolumeSpecName: "config-data") pod "d8900755-ba7e-4e00-9500-69afbdb69064" (UID: "d8900755-ba7e-4e00-9500-69afbdb69064"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295091 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295115 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295128 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8900755-ba7e-4e00-9500-69afbdb69064-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295139 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295151 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295161 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295172 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295171 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts" (OuterVolumeSpecName: "scripts") pod "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" (UID: "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.295187 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8900755-ba7e-4e00-9500-69afbdb69064-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.302957 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n" (OuterVolumeSpecName: "kube-api-access-bs79n") pod "d8900755-ba7e-4e00-9500-69afbdb69064" (UID: "d8900755-ba7e-4e00-9500-69afbdb69064"). InnerVolumeSpecName "kube-api-access-bs79n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.304086 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5" (OuterVolumeSpecName: "kube-api-access-x4jx5") pod "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" (UID: "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7"). InnerVolumeSpecName "kube-api-access-x4jx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.304141 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4" (OuterVolumeSpecName: "kube-api-access-cpcb4") pod "c58022e7-62a2-4ef5-bce1-024d8da0b9a8" (UID: "c58022e7-62a2-4ef5-bce1-024d8da0b9a8"). InnerVolumeSpecName "kube-api-access-cpcb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.305002 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc" (OuterVolumeSpecName: "kube-api-access-4g2fc") pod "ea4071c9-d190-4706-8f11-16b77f3db069" (UID: "ea4071c9-d190-4706-8f11-16b77f3db069"). InnerVolumeSpecName "kube-api-access-4g2fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.305832 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c58022e7-62a2-4ef5-bce1-024d8da0b9a8" (UID: "c58022e7-62a2-4ef5-bce1-024d8da0b9a8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.305876 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d8900755-ba7e-4e00-9500-69afbdb69064" (UID: "d8900755-ba7e-4e00-9500-69afbdb69064"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.305900 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" (UID: "fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.329819 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea4071c9-d190-4706-8f11-16b77f3db069" (UID: "ea4071c9-d190-4706-8f11-16b77f3db069"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.331392 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config" (OuterVolumeSpecName: "config") pod "ea4071c9-d190-4706-8f11-16b77f3db069" (UID: "ea4071c9-d190-4706-8f11-16b77f3db069"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.396492 4847 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.396843 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs79n\" (UniqueName: \"kubernetes.io/projected/d8900755-ba7e-4e00-9500-69afbdb69064-kube-api-access-bs79n\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.396960 4847 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d8900755-ba7e-4e00-9500-69afbdb69064-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397041 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpcb4\" (UniqueName: \"kubernetes.io/projected/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-kube-api-access-cpcb4\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397117 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397192 4847 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c58022e7-62a2-4ef5-bce1-024d8da0b9a8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397280 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397361 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4071c9-d190-4706-8f11-16b77f3db069-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397480 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g2fc\" (UniqueName: \"kubernetes.io/projected/ea4071c9-d190-4706-8f11-16b77f3db069-kube-api-access-4g2fc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.397585 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4jx5\" (UniqueName: \"kubernetes.io/projected/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7-kube-api-access-x4jx5\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.651974 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4f6b7576-wvhs4"] Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.783468 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f98d7b94c-594v6" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.783762 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f98d7b94c-594v6" event={"ID":"fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7","Type":"ContainerDied","Data":"ed736f92ebe1cad9d5bbe2c8f8de9de2291122113684f095d39d329432711448"} Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.788055 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f696c998f-hsmnv" event={"ID":"c58022e7-62a2-4ef5-bce1-024d8da0b9a8","Type":"ContainerDied","Data":"53dd00abfff2200847827b0c57311e05154de6f2c82e248f2d2c793475386a45"} Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.788117 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f696c998f-hsmnv" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.822275 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-66gqz" event={"ID":"ea4071c9-d190-4706-8f11-16b77f3db069","Type":"ContainerDied","Data":"50a93a0db345f9a038b26782d2e9368b214186e8abb90abad093db70bb0a39d4"} Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.822674 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a93a0db345f9a038b26782d2e9368b214186e8abb90abad093db70bb0a39d4" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.822789 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-66gqz" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.836386 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7985b45c4c-mt5vg" event={"ID":"d8900755-ba7e-4e00-9500-69afbdb69064","Type":"ContainerDied","Data":"6666dd22527733969080f6528ad6fe1ff475054cde5c0e9ab2dfae460bf49e9d"} Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.836490 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7985b45c4c-mt5vg" Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.909058 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.957198 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f98d7b94c-594v6"] Dec 10 14:43:11 crc kubenswrapper[4847]: I1210 14:43:11.996421 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.014454 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f696c998f-hsmnv"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.028564 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.036900 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7985b45c4c-mt5vg"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.049279 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.061643 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:12 crc kubenswrapper[4847]: E1210 14:43:12.062326 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" containerName="glance-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.062946 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" containerName="glance-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: E1210 14:43:12.063080 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="init" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.063165 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="init" Dec 10 14:43:12 crc kubenswrapper[4847]: E1210 14:43:12.063256 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4071c9-d190-4706-8f11-16b77f3db069" containerName="neutron-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.063331 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4071c9-d190-4706-8f11-16b77f3db069" containerName="neutron-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: E1210 14:43:12.063529 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.063611 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.064517 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" containerName="glance-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.064639 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea4071c9-d190-4706-8f11-16b77f3db069" containerName="neutron-db-sync" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.064763 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ce653b9-194a-4390-9a40-241ff10a18df" containerName="dnsmasq-dns" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.066091 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.078291 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116405 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116473 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116509 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsfkb\" (UniqueName: \"kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116570 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116603 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.116632 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.217953 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218108 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218148 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218183 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsfkb\" (UniqueName: \"kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218244 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.218989 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.222351 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.223521 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.223727 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.224029 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.243838 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsfkb\" (UniqueName: \"kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb\") pod \"dnsmasq-dns-56df8fb6b7-m7rmq\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.398367 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.536679 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.566644 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.573374 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.574172 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.608873 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.618490 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.624973 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.625013 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.624989 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.625235 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.625550 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dfjp4" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638016 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638082 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638129 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638161 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bdlz\" (UniqueName: \"kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638181 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.638302 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.739912 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps9lk\" (UniqueName: \"kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740016 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740067 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740111 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740282 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740366 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740390 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740442 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740475 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bdlz\" (UniqueName: \"kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740499 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.740903 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.741461 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.741582 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.741638 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.741867 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.742061 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.760258 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bdlz\" (UniqueName: \"kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz\") pod \"dnsmasq-dns-6b7b667979-6hsq9\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.773476 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58022e7-62a2-4ef5-bce1-024d8da0b9a8" path="/var/lib/kubelet/pods/c58022e7-62a2-4ef5-bce1-024d8da0b9a8/volumes" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.774144 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8900755-ba7e-4e00-9500-69afbdb69064" path="/var/lib/kubelet/pods/d8900755-ba7e-4e00-9500-69afbdb69064/volumes" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.774675 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7" path="/var/lib/kubelet/pods/fb18d2e2-1cc6-4fb2-98c6-9c13ba74f5f7/volumes" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.843825 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.843912 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.843962 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.844021 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps9lk\" (UniqueName: \"kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.844089 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.847659 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.847796 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.848243 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.848858 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.866792 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps9lk\" (UniqueName: \"kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk\") pod \"neutron-6c4f9cd7f4-7j84b\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.902735 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.904376 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.907014 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m498w" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.907227 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.907374 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.920261 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.929062 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945765 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945835 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9xbl\" (UniqueName: \"kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945859 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945886 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945920 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945973 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.945997 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:12 crc kubenswrapper[4847]: I1210 14:43:12.948193 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047409 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047479 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047546 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047571 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047620 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047668 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9xbl\" (UniqueName: \"kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.047687 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.048099 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.048537 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.049965 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.052256 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.053037 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.062058 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.065916 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9xbl\" (UniqueName: \"kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.081034 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.244076 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.257686 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.260232 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.264254 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.273757 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358765 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g9fl\" (UniqueName: \"kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358808 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358870 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358941 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358971 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.358992 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.359006 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461341 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461425 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g9fl\" (UniqueName: \"kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461532 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461619 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461676 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461742 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461763 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.461794 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.462422 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.462497 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.468157 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.468795 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.471234 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.480568 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g9fl\" (UniqueName: \"kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.502103 4847 scope.go:117] "RemoveContainer" containerID="d1849a7699352f6e2848c06dd0455ebc6b00e31bbe5af200d88b6de6faecba81" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.502915 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: E1210 14:43:13.512097 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 10 14:43:13 crc kubenswrapper[4847]: E1210 14:43:13.512238 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rg8k4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-twlv7_openstack(82ca7b48-ad83-46b4-a813-de1ec043a3e2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 14:43:13 crc kubenswrapper[4847]: E1210 14:43:13.513674 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-twlv7" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" Dec 10 14:43:13 crc kubenswrapper[4847]: W1210 14:43:13.521894 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7e7178d_cd26_4623_a05c_6eee95cb5dd6.slice/crio-dccdc187e6e5440fa70799a6eb027446cd27358be35b72b9bf919b969ef8697e WatchSource:0}: Error finding container dccdc187e6e5440fa70799a6eb027446cd27358be35b72b9bf919b969ef8697e: Status 404 returned error can't find the container with id dccdc187e6e5440fa70799a6eb027446cd27358be35b72b9bf919b969ef8697e Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.587025 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.652926 4847 scope.go:117] "RemoveContainer" containerID="2867c4dea42ce294c5d20ddba89e8bfd03cbf145a7c5ff9a6693b334d52af006" Dec 10 14:43:13 crc kubenswrapper[4847]: I1210 14:43:13.919886 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f6b7576-wvhs4" event={"ID":"f7e7178d-cd26-4623-a05c-6eee95cb5dd6","Type":"ContainerStarted","Data":"dccdc187e6e5440fa70799a6eb027446cd27358be35b72b9bf919b969ef8697e"} Dec 10 14:43:13 crc kubenswrapper[4847]: E1210 14:43:13.942162 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-twlv7" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.052488 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.068749 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:43:14 crc kubenswrapper[4847]: W1210 14:43:14.119841 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod517d4777_6273_401e_bf27_b23ea11f1a41.slice/crio-7a1ce065e6d8ac47fefffb536468e2b50c7ad3d69aae55948b0e351f522b65b3 WatchSource:0}: Error finding container 7a1ce065e6d8ac47fefffb536468e2b50c7ad3d69aae55948b0e351f522b65b3: Status 404 returned error can't find the container with id 7a1ce065e6d8ac47fefffb536468e2b50c7ad3d69aae55948b0e351f522b65b3 Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.208001 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fw2fp"] Dec 10 14:43:14 crc kubenswrapper[4847]: W1210 14:43:14.235819 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaac62992_6083_4fa3_bed2_587f24f6bb40.slice/crio-7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7 WatchSource:0}: Error finding container 7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7: Status 404 returned error can't find the container with id 7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7 Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.244900 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.948740 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f6b7576-wvhs4" event={"ID":"f7e7178d-cd26-4623-a05c-6eee95cb5dd6","Type":"ContainerStarted","Data":"7a733a87e475158b4da5a91ebfcf023739c11c9edac55f8eaaae1a20952d6aa6"} Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.967548 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187"} Dec 10 14:43:14 crc kubenswrapper[4847]: I1210 14:43:14.981138 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p6r7l" event={"ID":"66a6b89e-e44e-4e19-92bd-12d58ad99fd1","Type":"ContainerStarted","Data":"8413752fb972c003cf7769c652ddc5c3ea73dd479cd96ce2070628b29d74a6cb"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.005564 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerStarted","Data":"7a1ce065e6d8ac47fefffb536468e2b50c7ad3d69aae55948b0e351f522b65b3"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.024926 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j5hjt" event={"ID":"8e3440ea-24b3-4cc9-9872-8c4e28ab340b","Type":"ContainerStarted","Data":"48c20f73a3a36a202fdba3cdad451a92ddee0e3509aabeba2d8450409fbd3942"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.034580 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw2fp" event={"ID":"aac62992-6083-4fa3-bed2-587f24f6bb40","Type":"ContainerStarted","Data":"451ad0ad43f8dc82e00f2764d86a6ff59a5dabcc5536ca3c0a3f783a427cd911"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.034636 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw2fp" event={"ID":"aac62992-6083-4fa3-bed2-587f24f6bb40","Type":"ContainerStarted","Data":"7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.047647 4847 generic.go:334] "Generic (PLEG): container finished" podID="9efb7d3f-c254-429c-9cb2-68006df6f238" containerID="d2aa5f3dfa4c3f00359510868cd2e1febec78595d9dc719c703fc4880b479710" exitCode=0 Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.047696 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" event={"ID":"9efb7d3f-c254-429c-9cb2-68006df6f238","Type":"ContainerDied","Data":"d2aa5f3dfa4c3f00359510868cd2e1febec78595d9dc719c703fc4880b479710"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.047739 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" event={"ID":"9efb7d3f-c254-429c-9cb2-68006df6f238","Type":"ContainerStarted","Data":"c5f8e2e24749b7f959cfd570d3c7d909f0748f9cdd01575f871566b67270cdb3"} Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.065258 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-p6r7l" podStartSLOduration=7.305291321 podStartE2EDuration="38.065237638s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="2025-12-10 14:42:40.400109813 +0000 UTC m=+1109.969327433" lastFinishedPulling="2025-12-10 14:43:11.16005612 +0000 UTC m=+1140.729273750" observedRunningTime="2025-12-10 14:43:15.006169768 +0000 UTC m=+1144.575387398" watchObservedRunningTime="2025-12-10 14:43:15.065237638 +0000 UTC m=+1144.634455268" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.068285 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-j5hjt" podStartSLOduration=7.310739701 podStartE2EDuration="38.068272932s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="2025-12-10 14:42:40.400314038 +0000 UTC m=+1109.969531668" lastFinishedPulling="2025-12-10 14:43:11.157847259 +0000 UTC m=+1140.727064899" observedRunningTime="2025-12-10 14:43:15.045589196 +0000 UTC m=+1144.614806826" watchObservedRunningTime="2025-12-10 14:43:15.068272932 +0000 UTC m=+1144.637490562" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.101055 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fw2fp" podStartSLOduration=25.101037087 podStartE2EDuration="25.101037087s" podCreationTimestamp="2025-12-10 14:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:15.067056049 +0000 UTC m=+1144.636273679" watchObservedRunningTime="2025-12-10 14:43:15.101037087 +0000 UTC m=+1144.670254717" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.135983 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.320160 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.419596 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.482572 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.580201 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.651280 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739089 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739200 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-442qv\" (UniqueName: \"kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739230 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739264 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739493 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.739608 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc\") pod \"9efb7d3f-c254-429c-9cb2-68006df6f238\" (UID: \"9efb7d3f-c254-429c-9cb2-68006df6f238\") " Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.761873 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv" (OuterVolumeSpecName: "kube-api-access-442qv") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "kube-api-access-442qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.777216 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.814267 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.844672 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-442qv\" (UniqueName: \"kubernetes.io/projected/9efb7d3f-c254-429c-9cb2-68006df6f238-kube-api-access-442qv\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.844728 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.856357 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.856483 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.860383 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config" (OuterVolumeSpecName: "config") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.871313 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9efb7d3f-c254-429c-9cb2-68006df6f238" (UID: "9efb7d3f-c254-429c-9cb2-68006df6f238"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.946541 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.946914 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.946930 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:15 crc kubenswrapper[4847]: I1210 14:43:15.946939 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9efb7d3f-c254-429c-9cb2-68006df6f238-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.065159 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4f6b7576-wvhs4" event={"ID":"f7e7178d-cd26-4623-a05c-6eee95cb5dd6","Type":"ContainerStarted","Data":"ef2d2aa29da925833153b1a971b8a883bf6ea9a0d31824334c0973582f6a1ffc"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.070439 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerStarted","Data":"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.081540 4847 generic.go:334] "Generic (PLEG): container finished" podID="bd6f1587-775b-49a7-8b0e-39e5c074eda4" containerID="66141ddeba115f975809f7e9542b0a00eace4a49a8a8629acc6597db85202e00" exitCode=0 Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.081654 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" event={"ID":"bd6f1587-775b-49a7-8b0e-39e5c074eda4","Type":"ContainerDied","Data":"66141ddeba115f975809f7e9542b0a00eace4a49a8a8629acc6597db85202e00"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.081687 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" event={"ID":"bd6f1587-775b-49a7-8b0e-39e5c074eda4","Type":"ContainerStarted","Data":"30774c7d009c64c031dab94ace3b908b248da093d4ad2e0dd98d201dceb832ed"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.097076 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d4f6b7576-wvhs4" podStartSLOduration=29.326752535 podStartE2EDuration="30.097027834s" podCreationTimestamp="2025-12-10 14:42:46 +0000 UTC" firstStartedPulling="2025-12-10 14:43:13.531611431 +0000 UTC m=+1143.100829061" lastFinishedPulling="2025-12-10 14:43:14.30188673 +0000 UTC m=+1143.871104360" observedRunningTime="2025-12-10 14:43:16.089624291 +0000 UTC m=+1145.658841921" watchObservedRunningTime="2025-12-10 14:43:16.097027834 +0000 UTC m=+1145.666245484" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.097194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerStarted","Data":"95518f0b60049b75e8f6be5d82481be8b83b8881de27791f4602ba7bd588c02f"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.102073 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" event={"ID":"9efb7d3f-c254-429c-9cb2-68006df6f238","Type":"ContainerDied","Data":"c5f8e2e24749b7f959cfd570d3c7d909f0748f9cdd01575f871566b67270cdb3"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.102125 4847 scope.go:117] "RemoveContainer" containerID="d2aa5f3dfa4c3f00359510868cd2e1febec78595d9dc719c703fc4880b479710" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.102276 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-d7xkj" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.106603 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerStarted","Data":"0b69d80df901b676c4872b77cde7dd498a36852952319a21183bdd58c470130e"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.111197 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" event={"ID":"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8","Type":"ContainerStarted","Data":"95270f9527e693fc36c2156468f16d26f933181b391d2466c0b7ccfa2a5b537e"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.120558 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerStarted","Data":"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3"} Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.142741 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:16 crc kubenswrapper[4847]: W1210 14:43:16.195135 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded4ab9a4_a254_4aba_ab41_7b7a3e0eeb3d.slice/crio-95457e5c6ead1ff2deb361bc8c89566e7774e58b1ba8cc9975cc4dc5b6093c69 WatchSource:0}: Error finding container 95457e5c6ead1ff2deb361bc8c89566e7774e58b1ba8cc9975cc4dc5b6093c69: Status 404 returned error can't find the container with id 95457e5c6ead1ff2deb361bc8c89566e7774e58b1ba8cc9975cc4dc5b6093c69 Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.360078 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.395958 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-d7xkj"] Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.660758 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.661333 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:43:16 crc kubenswrapper[4847]: I1210 14:43:16.785238 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efb7d3f-c254-429c-9cb2-68006df6f238" path="/var/lib/kubelet/pods/9efb7d3f-c254-429c-9cb2-68006df6f238/volumes" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.008227 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076735 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076807 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076863 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076893 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsfkb\" (UniqueName: \"kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076912 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.076936 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config\") pod \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\" (UID: \"bd6f1587-775b-49a7-8b0e-39e5c074eda4\") " Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.135238 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb" (OuterVolumeSpecName: "kube-api-access-hsfkb") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "kube-api-access-hsfkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.139252 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.150204 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config" (OuterVolumeSpecName: "config") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.150821 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.167284 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerStarted","Data":"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.178721 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.178745 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsfkb\" (UniqueName: \"kubernetes.io/projected/bd6f1587-775b-49a7-8b0e-39e5c074eda4-kube-api-access-hsfkb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.178756 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.178766 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.181409 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" event={"ID":"bd6f1587-775b-49a7-8b0e-39e5c074eda4","Type":"ContainerDied","Data":"30774c7d009c64c031dab94ace3b908b248da093d4ad2e0dd98d201dceb832ed"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.181455 4847 scope.go:117] "RemoveContainer" containerID="66141ddeba115f975809f7e9542b0a00eace4a49a8a8629acc6597db85202e00" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.181585 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-m7rmq" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.188060 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.197195 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bd6f1587-775b-49a7-8b0e-39e5c074eda4" (UID: "bd6f1587-775b-49a7-8b0e-39e5c074eda4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.207931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerStarted","Data":"7a9278c853165bbb366d2b453ee252212864e89a1975e0e169e6fb53f88b09ca"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.213934 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8676865fc8-rvrhh" podStartSLOduration=30.317740917 podStartE2EDuration="31.21390519s" podCreationTimestamp="2025-12-10 14:42:46 +0000 UTC" firstStartedPulling="2025-12-10 14:43:14.125965185 +0000 UTC m=+1143.695182815" lastFinishedPulling="2025-12-10 14:43:15.022129458 +0000 UTC m=+1144.591347088" observedRunningTime="2025-12-10 14:43:17.20559062 +0000 UTC m=+1146.774808260" watchObservedRunningTime="2025-12-10 14:43:17.21390519 +0000 UTC m=+1146.783122840" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.256934 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c4f9cd7f4-7j84b" podStartSLOduration=5.256899486 podStartE2EDuration="5.256899486s" podCreationTimestamp="2025-12-10 14:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:17.252853105 +0000 UTC m=+1146.822070735" watchObservedRunningTime="2025-12-10 14:43:17.256899486 +0000 UTC m=+1146.826117116" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.258986 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerStarted","Data":"f699f7eec24505aa813fca8c96fdc344f55e8c142756a5b4e5267e97f6a61c42"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.269508 4847 generic.go:334] "Generic (PLEG): container finished" podID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerID="527171fd46b18e40948541c76cef5c6c38d973701d9bab2dcfd122cb3c8efa31" exitCode=0 Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.269591 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" event={"ID":"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8","Type":"ContainerDied","Data":"527171fd46b18e40948541c76cef5c6c38d973701d9bab2dcfd122cb3c8efa31"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.286337 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.286377 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd6f1587-775b-49a7-8b0e-39e5c074eda4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.289167 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerStarted","Data":"95457e5c6ead1ff2deb361bc8c89566e7774e58b1ba8cc9975cc4dc5b6093c69"} Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.564535 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:17 crc kubenswrapper[4847]: I1210 14:43:17.580919 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-m7rmq"] Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.308166 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerStarted","Data":"354a8d3c17fca29ecffa725ea38a20b3774996503a0f71e9e2ca6373d6267c4e"} Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.309978 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.320147 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" event={"ID":"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8","Type":"ContainerStarted","Data":"ef58603d18d1560101302959777cb2ad21ef5d3fe8abd79001769aa94efb0d2b"} Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.325053 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerStarted","Data":"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4"} Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.616393 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c86c8879c-drvf5"] Dec 10 14:43:18 crc kubenswrapper[4847]: E1210 14:43:18.617075 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efb7d3f-c254-429c-9cb2-68006df6f238" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.617103 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efb7d3f-c254-429c-9cb2-68006df6f238" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: E1210 14:43:18.617156 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd6f1587-775b-49a7-8b0e-39e5c074eda4" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.617163 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd6f1587-775b-49a7-8b0e-39e5c074eda4" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.617411 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd6f1587-775b-49a7-8b0e-39e5c074eda4" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.617453 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efb7d3f-c254-429c-9cb2-68006df6f238" containerName="init" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.618927 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.622385 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.627227 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.640573 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c86c8879c-drvf5"] Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745071 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745196 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-public-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745403 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-combined-ca-bundle\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745610 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-internal-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745732 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-httpd-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745767 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgsdb\" (UniqueName: \"kubernetes.io/projected/7d85a520-fbc1-4a46-b33e-7513646bfeaf-kube-api-access-dgsdb\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.745811 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-ovndb-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.862937 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-internal-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863047 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-httpd-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863074 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgsdb\" (UniqueName: \"kubernetes.io/projected/7d85a520-fbc1-4a46-b33e-7513646bfeaf-kube-api-access-dgsdb\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863114 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-ovndb-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863173 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863235 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-public-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.863338 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-combined-ca-bundle\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.885408 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-ovndb-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.892603 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-internal-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.893543 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.893863 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd6f1587-775b-49a7-8b0e-39e5c074eda4" path="/var/lib/kubelet/pods/bd6f1587-775b-49a7-8b0e-39e5c074eda4/volumes" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.901618 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-httpd-config\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.910162 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-public-tls-certs\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.910227 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d85a520-fbc1-4a46-b33e-7513646bfeaf-combined-ca-bundle\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:18 crc kubenswrapper[4847]: I1210 14:43:18.924787 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgsdb\" (UniqueName: \"kubernetes.io/projected/7d85a520-fbc1-4a46-b33e-7513646bfeaf-kube-api-access-dgsdb\") pod \"neutron-5c86c8879c-drvf5\" (UID: \"7d85a520-fbc1-4a46-b33e-7513646bfeaf\") " pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.063106 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.367916 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-log" containerID="cri-o://ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" gracePeriod=30 Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.368382 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-httpd" containerID="cri-o://7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" gracePeriod=30 Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.368397 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerStarted","Data":"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494"} Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.402941 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-log" containerID="cri-o://f699f7eec24505aa813fca8c96fdc344f55e8c142756a5b4e5267e97f6a61c42" gracePeriod=30 Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.403293 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-httpd" containerID="cri-o://5f973ee6b28a35022d3d8f66f5438529752a967472be55740de8ae55ecce5067" gracePeriod=30 Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.403445 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerStarted","Data":"5f973ee6b28a35022d3d8f66f5438529752a967472be55740de8ae55ecce5067"} Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.403551 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.416539 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.416515209 podStartE2EDuration="8.416515209s" podCreationTimestamp="2025-12-10 14:43:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:19.403544722 +0000 UTC m=+1148.972762352" watchObservedRunningTime="2025-12-10 14:43:19.416515209 +0000 UTC m=+1148.985732839" Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.497771 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" podStartSLOduration=7.497752682 podStartE2EDuration="7.497752682s" podCreationTimestamp="2025-12-10 14:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:19.430807774 +0000 UTC m=+1149.000025424" watchObservedRunningTime="2025-12-10 14:43:19.497752682 +0000 UTC m=+1149.066970312" Dec 10 14:43:19 crc kubenswrapper[4847]: I1210 14:43:19.518421 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.5184009320000005 podStartE2EDuration="7.518400932s" podCreationTimestamp="2025-12-10 14:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:19.457161232 +0000 UTC m=+1149.026378872" watchObservedRunningTime="2025-12-10 14:43:19.518400932 +0000 UTC m=+1149.087618562" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.001779 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c86c8879c-drvf5"] Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.309564 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428135 4847 generic.go:334] "Generic (PLEG): container finished" podID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerID="7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" exitCode=143 Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428637 4847 generic.go:334] "Generic (PLEG): container finished" podID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerID="ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" exitCode=143 Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428705 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerDied","Data":"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428769 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerDied","Data":"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428796 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d","Type":"ContainerDied","Data":"95457e5c6ead1ff2deb361bc8c89566e7774e58b1ba8cc9975cc4dc5b6093c69"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428816 4847 scope.go:117] "RemoveContainer" containerID="7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.428986 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.442949 4847 generic.go:334] "Generic (PLEG): container finished" podID="79328eac-25df-4e49-993d-5f2ff035be68" containerID="5f973ee6b28a35022d3d8f66f5438529752a967472be55740de8ae55ecce5067" exitCode=0 Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.442977 4847 generic.go:334] "Generic (PLEG): container finished" podID="79328eac-25df-4e49-993d-5f2ff035be68" containerID="f699f7eec24505aa813fca8c96fdc344f55e8c142756a5b4e5267e97f6a61c42" exitCode=143 Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.443030 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerDied","Data":"5f973ee6b28a35022d3d8f66f5438529752a967472be55740de8ae55ecce5067"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.443060 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerDied","Data":"f699f7eec24505aa813fca8c96fdc344f55e8c142756a5b4e5267e97f6a61c42"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.445634 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86c8879c-drvf5" event={"ID":"7d85a520-fbc1-4a46-b33e-7513646bfeaf","Type":"ContainerStarted","Data":"597d7e8e8796ba6b55041f750ca2d6ad75c4ee22eae600f69d0afa94e2fa7871"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.445725 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86c8879c-drvf5" event={"ID":"7d85a520-fbc1-4a46-b33e-7513646bfeaf","Type":"ContainerStarted","Data":"ef9e7b70fb7603a756d5d0d0412472ea82ce166c6437ac7225608bf493ca4f0a"} Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473637 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473756 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473835 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9xbl\" (UniqueName: \"kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473892 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473920 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.473982 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.474014 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs\") pod \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\" (UID: \"ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.475291 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.477251 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs" (OuterVolumeSpecName: "logs") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.480351 4847 scope.go:117] "RemoveContainer" containerID="ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.490354 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl" (OuterVolumeSpecName: "kube-api-access-j9xbl") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "kube-api-access-j9xbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.501302 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.504631 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts" (OuterVolumeSpecName: "scripts") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.543597 4847 scope.go:117] "RemoveContainer" containerID="7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.544415 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494\": container with ID starting with 7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494 not found: ID does not exist" containerID="7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.544446 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494"} err="failed to get container status \"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494\": rpc error: code = NotFound desc = could not find container \"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494\": container with ID starting with 7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494 not found: ID does not exist" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.544467 4847 scope.go:117] "RemoveContainer" containerID="ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.547029 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.556407 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4\": container with ID starting with ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4 not found: ID does not exist" containerID="ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.556453 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4"} err="failed to get container status \"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4\": rpc error: code = NotFound desc = could not find container \"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4\": container with ID starting with ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4 not found: ID does not exist" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.556478 4847 scope.go:117] "RemoveContainer" containerID="7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.561331 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494"} err="failed to get container status \"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494\": rpc error: code = NotFound desc = could not find container \"7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494\": container with ID starting with 7a83c3ed017615f33220947c12336792c319f2bab34bcb72e03f5af1f0bfe494 not found: ID does not exist" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.561372 4847 scope.go:117] "RemoveContainer" containerID="ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.565264 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4"} err="failed to get container status \"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4\": rpc error: code = NotFound desc = could not find container \"ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4\": container with ID starting with ceee594feeec656ca74267ce4bf4b712d6e0cf18feba12174ffaba53c43b3af4 not found: ID does not exist" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576255 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576285 4847 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576295 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9xbl\" (UniqueName: \"kubernetes.io/projected/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-kube-api-access-j9xbl\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576304 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576324 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.576333 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.609798 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data" (OuterVolumeSpecName: "config-data") pod "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" (UID: "ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.642554 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.677566 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.677600 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.716453 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.852186 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.874297 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889065 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.889497 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889515 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.889539 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889546 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.889558 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889564 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: E1210 14:43:20.889572 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889578 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889761 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889773 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-httpd" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889784 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="79328eac-25df-4e49-993d-5f2ff035be68" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.889796 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" containerName="glance-log" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890282 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890337 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g9fl\" (UniqueName: \"kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890409 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890457 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890494 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890586 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890636 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts\") pod \"79328eac-25df-4e49-993d-5f2ff035be68\" (UID: \"79328eac-25df-4e49-993d-5f2ff035be68\") " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.890653 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.892071 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.895765 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.897438 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.898269 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.898422 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.898925 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs" (OuterVolumeSpecName: "logs") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.905924 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts" (OuterVolumeSpecName: "scripts") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.911679 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl" (OuterVolumeSpecName: "kube-api-access-9g9fl") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "kube-api-access-9g9fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.961650 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992822 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992872 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992923 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992939 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992973 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.992988 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993015 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdmwr\" (UniqueName: \"kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993063 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993132 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993143 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g9fl\" (UniqueName: \"kubernetes.io/projected/79328eac-25df-4e49-993d-5f2ff035be68-kube-api-access-9g9fl\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993152 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993187 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993196 4847 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79328eac-25df-4e49-993d-5f2ff035be68-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:20 crc kubenswrapper[4847]: I1210 14:43:20.993205 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.046744 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.048226 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data" (OuterVolumeSpecName: "config-data") pod "79328eac-25df-4e49-993d-5f2ff035be68" (UID: "79328eac-25df-4e49-993d-5f2ff035be68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099137 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099298 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099355 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099691 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099734 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099801 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099823 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.099865 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdmwr\" (UniqueName: \"kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.100033 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79328eac-25df-4e49-993d-5f2ff035be68-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.100099 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.104546 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.104655 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.107242 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.112972 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.122084 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.125474 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdmwr\" (UniqueName: \"kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.126958 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.138664 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.216367 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.290846 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.524014 4847 generic.go:334] "Generic (PLEG): container finished" podID="66a6b89e-e44e-4e19-92bd-12d58ad99fd1" containerID="8413752fb972c003cf7769c652ddc5c3ea73dd479cd96ce2070628b29d74a6cb" exitCode=0 Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.524063 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p6r7l" event={"ID":"66a6b89e-e44e-4e19-92bd-12d58ad99fd1","Type":"ContainerDied","Data":"8413752fb972c003cf7769c652ddc5c3ea73dd479cd96ce2070628b29d74a6cb"} Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.551437 4847 generic.go:334] "Generic (PLEG): container finished" podID="8e3440ea-24b3-4cc9-9872-8c4e28ab340b" containerID="48c20f73a3a36a202fdba3cdad451a92ddee0e3509aabeba2d8450409fbd3942" exitCode=0 Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.551515 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j5hjt" event={"ID":"8e3440ea-24b3-4cc9-9872-8c4e28ab340b","Type":"ContainerDied","Data":"48c20f73a3a36a202fdba3cdad451a92ddee0e3509aabeba2d8450409fbd3942"} Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.562694 4847 generic.go:334] "Generic (PLEG): container finished" podID="aac62992-6083-4fa3-bed2-587f24f6bb40" containerID="451ad0ad43f8dc82e00f2764d86a6ff59a5dabcc5536ca3c0a3f783a427cd911" exitCode=0 Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.562763 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw2fp" event={"ID":"aac62992-6083-4fa3-bed2-587f24f6bb40","Type":"ContainerDied","Data":"451ad0ad43f8dc82e00f2764d86a6ff59a5dabcc5536ca3c0a3f783a427cd911"} Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.568101 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"79328eac-25df-4e49-993d-5f2ff035be68","Type":"ContainerDied","Data":"0b69d80df901b676c4872b77cde7dd498a36852952319a21183bdd58c470130e"} Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.568151 4847 scope.go:117] "RemoveContainer" containerID="5f973ee6b28a35022d3d8f66f5438529752a967472be55740de8ae55ecce5067" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.568264 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.593826 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86c8879c-drvf5" event={"ID":"7d85a520-fbc1-4a46-b33e-7513646bfeaf","Type":"ContainerStarted","Data":"81192cc5addf5c8ed24ed7a62fbd830a092d57192f8acf6397990cd0ca775c05"} Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.599846 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.685523 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c86c8879c-drvf5" podStartSLOduration=3.685159432 podStartE2EDuration="3.685159432s" podCreationTimestamp="2025-12-10 14:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:21.684984807 +0000 UTC m=+1151.254202447" watchObservedRunningTime="2025-12-10 14:43:21.685159432 +0000 UTC m=+1151.254377062" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.751329 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.818904 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.848177 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.849666 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.853227 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.866524 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 14:43:21 crc kubenswrapper[4847]: I1210 14:43:21.874330 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041488 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjnqj\" (UniqueName: \"kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041552 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041583 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041611 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041761 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041884 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.041932 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.042030 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.101934 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144386 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144614 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144655 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144783 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144867 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjnqj\" (UniqueName: \"kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144929 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.144982 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.145024 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.146375 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.146552 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.146585 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.154979 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.158862 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.164627 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.171280 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjnqj\" (UniqueName: \"kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.177271 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.183956 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.481249 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.976784 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79328eac-25df-4e49-993d-5f2ff035be68" path="/var/lib/kubelet/pods/79328eac-25df-4e49-993d-5f2ff035be68/volumes" Dec 10 14:43:22 crc kubenswrapper[4847]: I1210 14:43:22.977743 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d" path="/var/lib/kubelet/pods/ed4ab9a4-a254-4aba-ab41-7b7a3e0eeb3d/volumes" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.638652 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-j5hjt" event={"ID":"8e3440ea-24b3-4cc9-9872-8c4e28ab340b","Type":"ContainerDied","Data":"505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2"} Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.639058 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="505547ac41d5633685f1684eb2f57d11c5310ec85b3f476b23e888e53c690ff2" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.716779 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j5hjt" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.735050 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs\") pod \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.735138 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhvh6\" (UniqueName: \"kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6\") pod \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.735468 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data\") pod \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.735509 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts\") pod \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.735549 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle\") pod \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\" (UID: \"8e3440ea-24b3-4cc9-9872-8c4e28ab340b\") " Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.737643 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs" (OuterVolumeSpecName: "logs") pod "8e3440ea-24b3-4cc9-9872-8c4e28ab340b" (UID: "8e3440ea-24b3-4cc9-9872-8c4e28ab340b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.743622 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts" (OuterVolumeSpecName: "scripts") pod "8e3440ea-24b3-4cc9-9872-8c4e28ab340b" (UID: "8e3440ea-24b3-4cc9-9872-8c4e28ab340b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.758928 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6" (OuterVolumeSpecName: "kube-api-access-qhvh6") pod "8e3440ea-24b3-4cc9-9872-8c4e28ab340b" (UID: "8e3440ea-24b3-4cc9-9872-8c4e28ab340b"). InnerVolumeSpecName "kube-api-access-qhvh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.786988 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e3440ea-24b3-4cc9-9872-8c4e28ab340b" (UID: "8e3440ea-24b3-4cc9-9872-8c4e28ab340b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.823768 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data" (OuterVolumeSpecName: "config-data") pod "8e3440ea-24b3-4cc9-9872-8c4e28ab340b" (UID: "8e3440ea-24b3-4cc9-9872-8c4e28ab340b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.837378 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhvh6\" (UniqueName: \"kubernetes.io/projected/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-kube-api-access-qhvh6\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.837417 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.837428 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.837437 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:24 crc kubenswrapper[4847]: I1210 14:43:24.837446 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3440ea-24b3-4cc9-9872-8c4e28ab340b-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.648990 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-j5hjt" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.833373 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6d896bd648-rk9qk"] Dec 10 14:43:25 crc kubenswrapper[4847]: E1210 14:43:25.834329 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3440ea-24b3-4cc9-9872-8c4e28ab340b" containerName="placement-db-sync" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.834422 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3440ea-24b3-4cc9-9872-8c4e28ab340b" containerName="placement-db-sync" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.834748 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3440ea-24b3-4cc9-9872-8c4e28ab340b" containerName="placement-db-sync" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.835743 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.840648 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.840802 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.841148 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.841390 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.841769 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nf2pv" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.852661 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d896bd648-rk9qk"] Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.974338 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-internal-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.974628 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-config-data\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.974749 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-combined-ca-bundle\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.974878 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-scripts\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.974989 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7895251f-c0bb-40b7-a4a2-802a3517cd98-logs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.975080 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-public-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:25 crc kubenswrapper[4847]: I1210 14:43:25.975152 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z75n\" (UniqueName: \"kubernetes.io/projected/7895251f-c0bb-40b7-a4a2-802a3517cd98-kube-api-access-7z75n\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076757 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-scripts\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076816 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7895251f-c0bb-40b7-a4a2-802a3517cd98-logs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076850 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-public-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076880 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z75n\" (UniqueName: \"kubernetes.io/projected/7895251f-c0bb-40b7-a4a2-802a3517cd98-kube-api-access-7z75n\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076954 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-internal-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.076979 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-config-data\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.077044 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-combined-ca-bundle\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.077846 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7895251f-c0bb-40b7-a4a2-802a3517cd98-logs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.085629 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-internal-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.095085 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-combined-ca-bundle\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.095176 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-scripts\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.096497 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-config-data\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.098316 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7895251f-c0bb-40b7-a4a2-802a3517cd98-public-tls-certs\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.104505 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z75n\" (UniqueName: \"kubernetes.io/projected/7895251f-c0bb-40b7-a4a2-802a3517cd98-kube-api-access-7z75n\") pod \"placement-6d896bd648-rk9qk\" (UID: \"7895251f-c0bb-40b7-a4a2-802a3517cd98\") " pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.171174 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.532116 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.532261 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.533959 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 10 14:43:26 crc kubenswrapper[4847]: I1210 14:43:26.649593 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d4f6b7576-wvhs4" podUID="f7e7178d-cd26-4623-a05c-6eee95cb5dd6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 10 14:43:27 crc kubenswrapper[4847]: W1210 14:43:27.020781 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaccad8cb_ba38_44f0_bbac_20c879e8a2b9.slice/crio-0d57fec2d5f50462c6ffc403b82c3aa01781dc611e6524253b0fe8a999f991e9 WatchSource:0}: Error finding container 0d57fec2d5f50462c6ffc403b82c3aa01781dc611e6524253b0fe8a999f991e9: Status 404 returned error can't find the container with id 0d57fec2d5f50462c6ffc403b82c3aa01781dc611e6524253b0fe8a999f991e9 Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.079839 4847 scope.go:117] "RemoveContainer" containerID="f699f7eec24505aa813fca8c96fdc344f55e8c142756a5b4e5267e97f6a61c42" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.269083 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.315013 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.403959 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25x68\" (UniqueName: \"kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.404967 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data\") pod \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405020 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle\") pod \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405052 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405072 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405119 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405198 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405216 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmx4p\" (UniqueName: \"kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p\") pod \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\" (UID: \"66a6b89e-e44e-4e19-92bd-12d58ad99fd1\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.405250 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys\") pod \"aac62992-6083-4fa3-bed2-587f24f6bb40\" (UID: \"aac62992-6083-4fa3-bed2-587f24f6bb40\") " Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.410520 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68" (OuterVolumeSpecName: "kube-api-access-25x68") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "kube-api-access-25x68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.411104 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.414495 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "66a6b89e-e44e-4e19-92bd-12d58ad99fd1" (UID: "66a6b89e-e44e-4e19-92bd-12d58ad99fd1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.424335 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p" (OuterVolumeSpecName: "kube-api-access-xmx4p") pod "66a6b89e-e44e-4e19-92bd-12d58ad99fd1" (UID: "66a6b89e-e44e-4e19-92bd-12d58ad99fd1"). InnerVolumeSpecName "kube-api-access-xmx4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.435933 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts" (OuterVolumeSpecName: "scripts") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.454650 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.484857 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.502934 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66a6b89e-e44e-4e19-92bd-12d58ad99fd1" (UID: "66a6b89e-e44e-4e19-92bd-12d58ad99fd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512061 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25x68\" (UniqueName: \"kubernetes.io/projected/aac62992-6083-4fa3-bed2-587f24f6bb40-kube-api-access-25x68\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512092 4847 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512101 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512109 4847 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512117 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512126 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512136 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmx4p\" (UniqueName: \"kubernetes.io/projected/66a6b89e-e44e-4e19-92bd-12d58ad99fd1-kube-api-access-xmx4p\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.512144 4847 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.523961 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data" (OuterVolumeSpecName: "config-data") pod "aac62992-6083-4fa3-bed2-587f24f6bb40" (UID: "aac62992-6083-4fa3-bed2-587f24f6bb40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.613857 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac62992-6083-4fa3-bed2-587f24f6bb40-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.692774 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6d896bd648-rk9qk"] Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.694030 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p6r7l" event={"ID":"66a6b89e-e44e-4e19-92bd-12d58ad99fd1","Type":"ContainerDied","Data":"ed765f016f60dad6ce8022f4ba48e2f04dc240731f80f8bbe1382d00f7b028ef"} Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.694072 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed765f016f60dad6ce8022f4ba48e2f04dc240731f80f8bbe1382d00f7b028ef" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.694092 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p6r7l" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.699758 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerStarted","Data":"0d57fec2d5f50462c6ffc403b82c3aa01781dc611e6524253b0fe8a999f991e9"} Dec 10 14:43:27 crc kubenswrapper[4847]: W1210 14:43:27.700296 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7895251f_c0bb_40b7_a4a2_802a3517cd98.slice/crio-26e2f22943022a406f10c299a9e7eb621cbf7925227451e005e028d31b5dbe4b WatchSource:0}: Error finding container 26e2f22943022a406f10c299a9e7eb621cbf7925227451e005e028d31b5dbe4b: Status 404 returned error can't find the container with id 26e2f22943022a406f10c299a9e7eb621cbf7925227451e005e028d31b5dbe4b Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.707822 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw2fp" event={"ID":"aac62992-6083-4fa3-bed2-587f24f6bb40","Type":"ContainerDied","Data":"7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7"} Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.707873 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7770b6256ffbaf81107b4d676bff00c2e8f7690c1601f13a708144c2ceb524c7" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.708326 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw2fp" Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.719011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerStarted","Data":"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27"} Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.730561 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:43:27 crc kubenswrapper[4847]: I1210 14:43:27.924104 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.015236 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.016978 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="dnsmasq-dns" containerID="cri-o://9c52e790d767fb1767e08540f6c5f108fd6b5f0f456449f96dcc25484c438a42" gracePeriod=10 Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.073168 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.509624 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw"] Dec 10 14:43:28 crc kubenswrapper[4847]: E1210 14:43:28.531946 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac62992-6083-4fa3-bed2-587f24f6bb40" containerName="keystone-bootstrap" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.531963 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac62992-6083-4fa3-bed2-587f24f6bb40" containerName="keystone-bootstrap" Dec 10 14:43:28 crc kubenswrapper[4847]: E1210 14:43:28.531991 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a6b89e-e44e-4e19-92bd-12d58ad99fd1" containerName="barbican-db-sync" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.531998 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a6b89e-e44e-4e19-92bd-12d58ad99fd1" containerName="barbican-db-sync" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.532165 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="66a6b89e-e44e-4e19-92bd-12d58ad99fd1" containerName="barbican-db-sync" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.532176 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac62992-6083-4fa3-bed2-587f24f6bb40" containerName="keystone-bootstrap" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.545874 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.547959 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-combined-ca-bundle\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.548010 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.548134 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70fb729f-c877-4d76-a5b0-a84470c6c4bc-logs\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.548172 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szrls\" (UniqueName: \"kubernetes.io/projected/70fb729f-c877-4d76-a5b0-a84470c6c4bc-kube-api-access-szrls\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.548222 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data-custom\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.551142 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.551382 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rlqtm" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.570341 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.584776 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.590929 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-54599d665c-kl29l"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.592436 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.598627 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.598939 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.599188 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.599366 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mz28n" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.599511 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.600570 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.630779 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-b85c87b55-fkkkc"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.632647 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.663698 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70fb729f-c877-4d76-a5b0-a84470c6c4bc-logs\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.663771 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szrls\" (UniqueName: \"kubernetes.io/projected/70fb729f-c877-4d76-a5b0-a84470c6c4bc-kube-api-access-szrls\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.663815 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data-custom\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.663860 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-combined-ca-bundle\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.663894 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.664804 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.665281 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70fb729f-c877-4d76-a5b0-a84470c6c4bc-logs\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.670513 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54599d665c-kl29l"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.676424 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.683563 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-combined-ca-bundle\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.691910 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70fb729f-c877-4d76-a5b0-a84470c6c4bc-config-data-custom\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.704304 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szrls\" (UniqueName: \"kubernetes.io/projected/70fb729f-c877-4d76-a5b0-a84470c6c4bc-kube-api-access-szrls\") pod \"barbican-keystone-listener-9fcb8ccdb-j6kdw\" (UID: \"70fb729f-c877-4d76-a5b0-a84470c6c4bc\") " pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.741777 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b85c87b55-fkkkc"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.767500 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-credential-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.767601 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770176 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqg4p\" (UniqueName: \"kubernetes.io/projected/9e5b990d-5058-4827-846b-e5ba776fa0ba-kube-api-access-gqg4p\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770282 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-config-data\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770355 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-combined-ca-bundle\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770483 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data-custom\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770510 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlfq4\" (UniqueName: \"kubernetes.io/projected/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-kube-api-access-qlfq4\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770545 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-scripts\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770618 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-public-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770645 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-fernet-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770695 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-internal-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770920 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-combined-ca-bundle\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.770948 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e5b990d-5058-4827-846b-e5ba776fa0ba-logs\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.795684 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.798467 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerStarted","Data":"5e55074c6e6d6f862dbb3be9c6bc04a1b178630bb58a096302288e4503ca500c"} Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.798639 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.809924 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.820196 4847 generic.go:334] "Generic (PLEG): container finished" podID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerID="9c52e790d767fb1767e08540f6c5f108fd6b5f0f456449f96dcc25484c438a42" exitCode=0 Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.820312 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" event={"ID":"cfa565dd-ab76-4e5b-a464-369a60ccd1d3","Type":"ContainerDied","Data":"9c52e790d767fb1767e08540f6c5f108fd6b5f0f456449f96dcc25484c438a42"} Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.850683 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerStarted","Data":"2639bd1f20c0ca05880a97984295b32c9c4bc7f6247346082a23c53bd2718b60"} Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875411 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-combined-ca-bundle\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875453 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e5b990d-5058-4827-846b-e5ba776fa0ba-logs\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875484 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875521 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-credential-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875554 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875572 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875570 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d896bd648-rk9qk" event={"ID":"7895251f-c0bb-40b7-a4a2-802a3517cd98","Type":"ContainerStarted","Data":"2d76f2b36d2c15f7d46acdc4fde9f6794f7be10805d9c0ac2d4be1e6daca574d"} Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875610 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d896bd648-rk9qk" event={"ID":"7895251f-c0bb-40b7-a4a2-802a3517cd98","Type":"ContainerStarted","Data":"26e2f22943022a406f10c299a9e7eb621cbf7925227451e005e028d31b5dbe4b"} Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875590 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqg4p\" (UniqueName: \"kubernetes.io/projected/9e5b990d-5058-4827-846b-e5ba776fa0ba-kube-api-access-gqg4p\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875701 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-config-data\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875768 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-combined-ca-bundle\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875819 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875937 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data-custom\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875968 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlfq4\" (UniqueName: \"kubernetes.io/projected/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-kube-api-access-qlfq4\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.875990 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-scripts\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876050 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876127 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-public-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876148 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-fernet-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876174 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-internal-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876193 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlxxz\" (UniqueName: \"kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.876245 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.884556 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e5b990d-5058-4827-846b-e5ba776fa0ba-logs\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.885851 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-credential-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.886802 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data-custom\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.887506 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-config-data\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.890629 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-combined-ca-bundle\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.891762 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e5b990d-5058-4827-846b-e5ba776fa0ba-config-data\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.902854 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.905410 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.905562 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-internal-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.906540 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-scripts\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.906601 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-public-tls-certs\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.907474 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-fernet-keys\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.916378 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.916592 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlfq4\" (UniqueName: \"kubernetes.io/projected/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-kube-api-access-qlfq4\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.916774 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.922472 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f2f4d1f-1fc9-48cd-8697-bdc788483a1c-combined-ca-bundle\") pod \"keystone-54599d665c-kl29l\" (UID: \"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c\") " pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.922520 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqg4p\" (UniqueName: \"kubernetes.io/projected/9e5b990d-5058-4827-846b-e5ba776fa0ba-kube-api-access-gqg4p\") pod \"barbican-worker-b85c87b55-fkkkc\" (UID: \"9e5b990d-5058-4827-846b-e5ba776fa0ba\") " pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.980876 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.980955 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzndd\" (UniqueName: \"kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981003 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981038 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlxxz\" (UniqueName: \"kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981062 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981567 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981693 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981778 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.981954 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.982126 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.982153 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.982240 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.982846 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.983495 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.983589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:28 crc kubenswrapper[4847]: I1210 14:43:28.984246 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.002795 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlxxz\" (UniqueName: \"kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz\") pod \"dnsmasq-dns-848cf88cfc-bwcb9\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.017482 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.083850 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.083938 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.083985 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzndd\" (UniqueName: \"kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.084034 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.084098 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.085215 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.089313 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.089517 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.089573 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.099632 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.104974 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzndd\" (UniqueName: \"kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd\") pod \"barbican-api-7587c5f9d-h9fj2\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.129890 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-b85c87b55-fkkkc" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.188585 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.247752 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.940011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-twlv7" event={"ID":"82ca7b48-ad83-46b4-a813-de1ec043a3e2","Type":"ContainerStarted","Data":"3e334ccf765e497d8ace6e292f05aef56b61a77a92433f93f0064cc662088c75"} Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.995994 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6d896bd648-rk9qk" event={"ID":"7895251f-c0bb-40b7-a4a2-802a3517cd98","Type":"ContainerStarted","Data":"5fff25721b15a5338b09bcd5a441d7196e14fdffaed16f85e04588b09a1f8cd1"} Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.997546 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:29 crc kubenswrapper[4847]: I1210 14:43:29.997579 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.016665 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-twlv7" podStartSLOduration=4.5928226599999995 podStartE2EDuration="53.016650374s" podCreationTimestamp="2025-12-10 14:42:37 +0000 UTC" firstStartedPulling="2025-12-10 14:42:39.05435106 +0000 UTC m=+1108.623568690" lastFinishedPulling="2025-12-10 14:43:27.478178774 +0000 UTC m=+1157.047396404" observedRunningTime="2025-12-10 14:43:30.012622043 +0000 UTC m=+1159.581839673" watchObservedRunningTime="2025-12-10 14:43:30.016650374 +0000 UTC m=+1159.585868004" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.059269 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerStarted","Data":"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640"} Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.070229 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6d896bd648-rk9qk" podStartSLOduration=5.070205182 podStartE2EDuration="5.070205182s" podCreationTimestamp="2025-12-10 14:43:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:30.053910112 +0000 UTC m=+1159.623127742" watchObservedRunningTime="2025-12-10 14:43:30.070205182 +0000 UTC m=+1159.639422812" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.172126 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.220750 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54599d665c-kl29l"] Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.254450 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config\") pod \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.254521 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2dvg\" (UniqueName: \"kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg\") pod \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.264457 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc\") pod \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.264520 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb\") pod \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.264555 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb\") pod \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\" (UID: \"cfa565dd-ab76-4e5b-a464-369a60ccd1d3\") " Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.279939 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg" (OuterVolumeSpecName: "kube-api-access-g2dvg") pod "cfa565dd-ab76-4e5b-a464-369a60ccd1d3" (UID: "cfa565dd-ab76-4e5b-a464-369a60ccd1d3"). InnerVolumeSpecName "kube-api-access-g2dvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.369384 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2dvg\" (UniqueName: \"kubernetes.io/projected/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-kube-api-access-g2dvg\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.372671 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cfa565dd-ab76-4e5b-a464-369a60ccd1d3" (UID: "cfa565dd-ab76-4e5b-a464-369a60ccd1d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.409197 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config" (OuterVolumeSpecName: "config") pod "cfa565dd-ab76-4e5b-a464-369a60ccd1d3" (UID: "cfa565dd-ab76-4e5b-a464-369a60ccd1d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.432830 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cfa565dd-ab76-4e5b-a464-369a60ccd1d3" (UID: "cfa565dd-ab76-4e5b-a464-369a60ccd1d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.436514 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cfa565dd-ab76-4e5b-a464-369a60ccd1d3" (UID: "cfa565dd-ab76-4e5b-a464-369a60ccd1d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.470540 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.470566 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.470580 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.470604 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfa565dd-ab76-4e5b-a464-369a60ccd1d3-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.680112 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.709327 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.719817 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw"] Dec 10 14:43:30 crc kubenswrapper[4847]: I1210 14:43:30.824930 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-b85c87b55-fkkkc"] Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.075989 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" event={"ID":"3152f661-df13-4d77-a736-ddf382434a2c","Type":"ContainerStarted","Data":"e3718bec7015c83b6066ae1624bf1a5b542d6a3fc4e5cf6759e6501dc0b0b649"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.077929 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" event={"ID":"70fb729f-c877-4d76-a5b0-a84470c6c4bc","Type":"ContainerStarted","Data":"fb404d0902201878e5f516d2d1c3237bfb3a6c6207dbb7dce32f9c7a709e9c2d"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.087602 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerStarted","Data":"7ad994b83d9fff91f53501bf519bc01d2288f7d2c3cae9c275388ff9ae8cf73c"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.092423 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b85c87b55-fkkkc" event={"ID":"9e5b990d-5058-4827-846b-e5ba776fa0ba","Type":"ContainerStarted","Data":"56fac69c631e082ef85191bf361738aa98c8b0bb0942225ca3b6b0d57f78c882"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.095922 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54599d665c-kl29l" event={"ID":"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c","Type":"ContainerStarted","Data":"ee94e612078712b9abbf5ddeb995070781e80f026b6f13daeb1e10fb3276e54c"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.095996 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54599d665c-kl29l" event={"ID":"8f2f4d1f-1fc9-48cd-8697-bdc788483a1c","Type":"ContainerStarted","Data":"e2a69544e9bf42f39435d7a7620568dbde74803fc06bbe55d2747ccb33dee364"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.096065 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.101679 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.101826 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-57vhg" event={"ID":"cfa565dd-ab76-4e5b-a464-369a60ccd1d3","Type":"ContainerDied","Data":"6f584d4636d43a5d94dcc25d9d6ca3f3a1387d1eddf11b57321f0177d9348ded"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.101919 4847 scope.go:117] "RemoveContainer" containerID="9c52e790d767fb1767e08540f6c5f108fd6b5f0f456449f96dcc25484c438a42" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.109485 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerStarted","Data":"109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729"} Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.135067 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-54599d665c-kl29l" podStartSLOduration=3.13504985 podStartE2EDuration="3.13504985s" podCreationTimestamp="2025-12-10 14:43:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:31.122539155 +0000 UTC m=+1160.691756785" watchObservedRunningTime="2025-12-10 14:43:31.13504985 +0000 UTC m=+1160.704267480" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.155991 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.155944368 podStartE2EDuration="11.155944368s" podCreationTimestamp="2025-12-10 14:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:31.150552079 +0000 UTC m=+1160.719769709" watchObservedRunningTime="2025-12-10 14:43:31.155944368 +0000 UTC m=+1160.725161998" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.299192 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.299745 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.310018 4847 scope.go:117] "RemoveContainer" containerID="495c457f7386b14df806c517ee4578323d1bf0eee51d60f631890d0a1c759fbc" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.331886 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.357458 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-57vhg"] Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.360243 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.375769 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.861105 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5488ccdfb4-gmxkx"] Dec 10 14:43:31 crc kubenswrapper[4847]: E1210 14:43:31.861882 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="dnsmasq-dns" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.861907 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="dnsmasq-dns" Dec 10 14:43:31 crc kubenswrapper[4847]: E1210 14:43:31.861930 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="init" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.861936 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="init" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.862096 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" containerName="dnsmasq-dns" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.865329 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.868021 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.868216 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 10 14:43:31 crc kubenswrapper[4847]: I1210 14:43:31.877023 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5488ccdfb4-gmxkx"] Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.008753 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-internal-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.008871 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8pqk\" (UniqueName: \"kubernetes.io/projected/e4d02170-155d-4380-9e77-e8705b47a968-kube-api-access-q8pqk\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.009023 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.009060 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-combined-ca-bundle\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.009099 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d02170-155d-4380-9e77-e8705b47a968-logs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.009121 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-public-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.009354 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data-custom\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.111498 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data-custom\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.111607 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-internal-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.112035 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8pqk\" (UniqueName: \"kubernetes.io/projected/e4d02170-155d-4380-9e77-e8705b47a968-kube-api-access-q8pqk\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.112139 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.112168 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-combined-ca-bundle\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.112211 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d02170-155d-4380-9e77-e8705b47a968-logs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.112239 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-public-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.113044 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d02170-155d-4380-9e77-e8705b47a968-logs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.146084 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-public-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.148869 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerStarted","Data":"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1"} Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.150236 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-internal-tls-certs\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.154089 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data-custom\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.157577 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-config-data\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.162326 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8pqk\" (UniqueName: \"kubernetes.io/projected/e4d02170-155d-4380-9e77-e8705b47a968-kube-api-access-q8pqk\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.171646 4847 generic.go:334] "Generic (PLEG): container finished" podID="3152f661-df13-4d77-a736-ddf382434a2c" containerID="f60f1ccd2e32d8c810ba0ee6bccd1dbcbda256bd8286ff17a53458239efe52ef" exitCode=0 Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.171741 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" event={"ID":"3152f661-df13-4d77-a736-ddf382434a2c","Type":"ContainerDied","Data":"f60f1ccd2e32d8c810ba0ee6bccd1dbcbda256bd8286ff17a53458239efe52ef"} Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.181786 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerStarted","Data":"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421"} Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.181846 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerStarted","Data":"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67"} Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.183268 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.183415 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.183511 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.187565 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d02170-155d-4380-9e77-e8705b47a968-combined-ca-bundle\") pod \"barbican-api-5488ccdfb4-gmxkx\" (UID: \"e4d02170-155d-4380-9e77-e8705b47a968\") " pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.210872 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.231176 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.231151112 podStartE2EDuration="11.231151112s" podCreationTimestamp="2025-12-10 14:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:32.207195791 +0000 UTC m=+1161.776413431" watchObservedRunningTime="2025-12-10 14:43:32.231151112 +0000 UTC m=+1161.800368742" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.280822 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7587c5f9d-h9fj2" podStartSLOduration=4.280779381 podStartE2EDuration="4.280779381s" podCreationTimestamp="2025-12-10 14:43:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:32.25574869 +0000 UTC m=+1161.824966330" watchObservedRunningTime="2025-12-10 14:43:32.280779381 +0000 UTC m=+1161.849997011" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.482040 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.482399 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.541976 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.555449 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.653677 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5488ccdfb4-gmxkx"] Dec 10 14:43:32 crc kubenswrapper[4847]: I1210 14:43:32.780212 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa565dd-ab76-4e5b-a464-369a60ccd1d3" path="/var/lib/kubelet/pods/cfa565dd-ab76-4e5b-a464-369a60ccd1d3/volumes" Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.198067 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" event={"ID":"3152f661-df13-4d77-a736-ddf382434a2c","Type":"ContainerStarted","Data":"c940e57e6381a801026597659fc6417f1b1fc8005c80ec459db203abdbc7599d"} Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.199047 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.201981 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5488ccdfb4-gmxkx" event={"ID":"e4d02170-155d-4380-9e77-e8705b47a968","Type":"ContainerStarted","Data":"9a4c7638c35bf597b09f2f92b580b77b9d5cc70ddde8d3df5609a420dfcf782b"} Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.203109 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.203171 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.203193 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:33 crc kubenswrapper[4847]: I1210 14:43:33.220366 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" podStartSLOduration=5.220338203 podStartE2EDuration="5.220338203s" podCreationTimestamp="2025-12-10 14:43:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:33.218854232 +0000 UTC m=+1162.788071862" watchObservedRunningTime="2025-12-10 14:43:33.220338203 +0000 UTC m=+1162.789555833" Dec 10 14:43:34 crc kubenswrapper[4847]: I1210 14:43:34.222455 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5488ccdfb4-gmxkx" event={"ID":"e4d02170-155d-4380-9e77-e8705b47a968","Type":"ContainerStarted","Data":"7d24116177e7aae98104ad3912467902c59810e0197f1488f139f5a4cb88e201"} Dec 10 14:43:35 crc kubenswrapper[4847]: I1210 14:43:35.319975 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 14:43:36 crc kubenswrapper[4847]: I1210 14:43:36.426654 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:36 crc kubenswrapper[4847]: I1210 14:43:36.532573 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 10 14:43:36 crc kubenswrapper[4847]: I1210 14:43:36.647643 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d4f6b7576-wvhs4" podUID="f7e7178d-cd26-4623-a05c-6eee95cb5dd6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 10 14:43:37 crc kubenswrapper[4847]: I1210 14:43:37.259658 4847 generic.go:334] "Generic (PLEG): container finished" podID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" containerID="3e334ccf765e497d8ace6e292f05aef56b61a77a92433f93f0064cc662088c75" exitCode=0 Dec 10 14:43:37 crc kubenswrapper[4847]: I1210 14:43:37.259733 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-twlv7" event={"ID":"82ca7b48-ad83-46b4-a813-de1ec043a3e2","Type":"ContainerDied","Data":"3e334ccf765e497d8ace6e292f05aef56b61a77a92433f93f0064cc662088c75"} Dec 10 14:43:37 crc kubenswrapper[4847]: I1210 14:43:37.439051 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.535701 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.753233 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-twlv7" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878481 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878634 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878689 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878742 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878873 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg8k4\" (UniqueName: \"kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.878936 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle\") pod \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\" (UID: \"82ca7b48-ad83-46b4-a813-de1ec043a3e2\") " Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.879677 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.887907 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4" (OuterVolumeSpecName: "kube-api-access-rg8k4") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "kube-api-access-rg8k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.887995 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts" (OuterVolumeSpecName: "scripts") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.902868 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.930835 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.970266 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data" (OuterVolumeSpecName: "config-data") pod "82ca7b48-ad83-46b4-a813-de1ec043a3e2" (UID: "82ca7b48-ad83-46b4-a813-de1ec043a3e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981057 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg8k4\" (UniqueName: \"kubernetes.io/projected/82ca7b48-ad83-46b4-a813-de1ec043a3e2-kube-api-access-rg8k4\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981094 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981109 4847 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ca7b48-ad83-46b4-a813-de1ec043a3e2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981121 4847 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981132 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:38 crc kubenswrapper[4847]: I1210 14:43:38.981199 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ca7b48-ad83-46b4-a813-de1ec043a3e2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.191875 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.274841 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.275314 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="dnsmasq-dns" containerID="cri-o://ef58603d18d1560101302959777cb2ad21ef5d3fe8abd79001769aa94efb0d2b" gracePeriod=10 Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.309177 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-twlv7" event={"ID":"82ca7b48-ad83-46b4-a813-de1ec043a3e2","Type":"ContainerDied","Data":"97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176"} Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.309231 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97ecce7f04f0ae5b15f5dc953ae913229da0e4084a4e126b50e1a6546fa6c176" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.309319 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-twlv7" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.653948 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:39 crc kubenswrapper[4847]: E1210 14:43:39.654275 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" containerName="cinder-db-sync" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.654287 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" containerName="cinder-db-sync" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.654534 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" containerName="cinder-db-sync" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.655436 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.657039 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.660997 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-blk76" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.661214 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.661346 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.713651 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.738179 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.742519 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.779785 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810027 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810081 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810199 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810215 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhnvw\" (UniqueName: \"kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810341 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.810360 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.905587 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.907463 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911738 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911793 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911829 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npmt9\" (UniqueName: \"kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911853 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911873 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911894 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911948 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911968 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.911992 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.912069 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.912090 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhnvw\" (UniqueName: \"kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.912124 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.915009 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.921589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.924078 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.927856 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.929227 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.930738 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.936280 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.945958 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhnvw\" (UniqueName: \"kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw\") pod \"cinder-scheduler-0\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:39 crc kubenswrapper[4847]: I1210 14:43:39.987796 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.013381 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.013422 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.013448 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019432 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019535 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019573 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdnld\" (UniqueName: \"kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019607 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019633 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019690 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npmt9\" (UniqueName: \"kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019705 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019775 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019802 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.019906 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.020861 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.020959 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.021640 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.025324 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.026690 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.040839 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npmt9\" (UniqueName: \"kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9\") pod \"dnsmasq-dns-6578955fd5-q4q79\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.095695 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.131733 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132057 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132209 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132277 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132344 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132445 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdnld\" (UniqueName: \"kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.132520 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.141739 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.143249 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.151349 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.151425 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.151852 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.155452 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.186176 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdnld\" (UniqueName: \"kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld\") pod \"cinder-api-0\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.325729 4847 generic.go:334] "Generic (PLEG): container finished" podID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerID="ef58603d18d1560101302959777cb2ad21ef5d3fe8abd79001769aa94efb0d2b" exitCode=0 Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.325785 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" event={"ID":"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8","Type":"ContainerDied","Data":"ef58603d18d1560101302959777cb2ad21ef5d3fe8abd79001769aa94efb0d2b"} Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.417059 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:40 crc kubenswrapper[4847]: I1210 14:43:40.826131 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7587c5f9d-h9fj2" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 14:43:41 crc kubenswrapper[4847]: I1210 14:43:41.226002 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:41 crc kubenswrapper[4847]: I1210 14:43:41.574474 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:41 crc kubenswrapper[4847]: I1210 14:43:41.944497 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.369203 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" event={"ID":"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8","Type":"ContainerDied","Data":"95270f9527e693fc36c2156468f16d26f933181b391d2466c0b7ccfa2a5b537e"} Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.369487 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95270f9527e693fc36c2156468f16d26f933181b391d2466c0b7ccfa2a5b537e" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.374301 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.489411 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.489827 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bdlz\" (UniqueName: \"kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.489863 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.489914 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.490029 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.490097 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb\") pod \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\" (UID: \"02e9d21c-37e3-40c4-83dc-c2a00b9df7f8\") " Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.510018 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz" (OuterVolumeSpecName: "kube-api-access-2bdlz") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "kube-api-access-2bdlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.592002 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bdlz\" (UniqueName: \"kubernetes.io/projected/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-kube-api-access-2bdlz\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.641405 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.690642 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.694509 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.694543 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.722660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.768150 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config" (OuterVolumeSpecName: "config") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.796173 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.796199 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.806009 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" (UID: "02e9d21c-37e3-40c4-83dc-c2a00b9df7f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.851357 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:42 crc kubenswrapper[4847]: E1210 14:43:42.890113 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.900738 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:42 crc kubenswrapper[4847]: I1210 14:43:42.976082 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.125088 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.158160 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.420525 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerStarted","Data":"5c79f214ba2cee819c9b03c36a943ed310811aa0bbd4ff5047e5ddab1dbd91ef"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.423991 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerStarted","Data":"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.424158 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="ceilometer-notification-agent" containerID="cri-o://062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3" gracePeriod=30 Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.424417 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.424584 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="proxy-httpd" containerID="cri-o://8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4" gracePeriod=30 Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.424581 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="sg-core" containerID="cri-o://58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27" gracePeriod=30 Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.431578 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerStarted","Data":"e9b7afe6e85fc31d8a000795732034193c7b52ebaa895471d0636583ff9957a1"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.434578 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b85c87b55-fkkkc" event={"ID":"9e5b990d-5058-4827-846b-e5ba776fa0ba","Type":"ContainerStarted","Data":"a4d4efbb41bc37bb52a322d59477c1e9da95bb01092398236f9d6b0302961e90"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.435385 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-b85c87b55-fkkkc" event={"ID":"9e5b990d-5058-4827-846b-e5ba776fa0ba","Type":"ContainerStarted","Data":"4517c6bd52c7844abe7d0d12cc99101f56f31eb84536bef1628b5c8c1d216012"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.439079 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" event={"ID":"159c593e-de1b-4dfe-b846-d0da68a0881a","Type":"ContainerStarted","Data":"fe148ffc66510b6b23f27bcf718f797926843935590a6947f969e258b90699d2"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.447982 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" event={"ID":"70fb729f-c877-4d76-a5b0-a84470c6c4bc","Type":"ContainerStarted","Data":"77ab51566a8bf5ffed660437e1cb359b0a38c527e785145d98eef03c6f58e409"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.448197 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" event={"ID":"70fb729f-c877-4d76-a5b0-a84470c6c4bc","Type":"ContainerStarted","Data":"6e2a65b8ca727ea8087cedcbe70779c7e5a80453c6e126e67a562bd55de4a7f5"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.454690 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6hsq9" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.455895 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5488ccdfb4-gmxkx" event={"ID":"e4d02170-155d-4380-9e77-e8705b47a968","Type":"ContainerStarted","Data":"9e16ec1363c2372ef7537b0159a78d67ca5cf548e95c59e402c4997bf4a16ecc"} Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.455947 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.455959 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.476369 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-b85c87b55-fkkkc" podStartSLOduration=4.169256975 podStartE2EDuration="15.47634395s" podCreationTimestamp="2025-12-10 14:43:28 +0000 UTC" firstStartedPulling="2025-12-10 14:43:30.839976347 +0000 UTC m=+1160.409193977" lastFinishedPulling="2025-12-10 14:43:42.147063322 +0000 UTC m=+1171.716280952" observedRunningTime="2025-12-10 14:43:43.464953185 +0000 UTC m=+1173.034170815" watchObservedRunningTime="2025-12-10 14:43:43.47634395 +0000 UTC m=+1173.045561600" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.520318 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-9fcb8ccdb-j6kdw" podStartSLOduration=4.135727208 podStartE2EDuration="15.520295572s" podCreationTimestamp="2025-12-10 14:43:28 +0000 UTC" firstStartedPulling="2025-12-10 14:43:30.766415117 +0000 UTC m=+1160.335632747" lastFinishedPulling="2025-12-10 14:43:42.150983471 +0000 UTC m=+1171.720201111" observedRunningTime="2025-12-10 14:43:43.486917011 +0000 UTC m=+1173.056134641" watchObservedRunningTime="2025-12-10 14:43:43.520295572 +0000 UTC m=+1173.089513202" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.563596 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5488ccdfb4-gmxkx" podStartSLOduration=12.563574777 podStartE2EDuration="12.563574777s" podCreationTimestamp="2025-12-10 14:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:43.511758616 +0000 UTC m=+1173.080976257" watchObservedRunningTime="2025-12-10 14:43:43.563574777 +0000 UTC m=+1173.132792407" Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.578983 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:43 crc kubenswrapper[4847]: I1210 14:43:43.587687 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6hsq9"] Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.479657 4847 generic.go:334] "Generic (PLEG): container finished" podID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerID="8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4" exitCode=0 Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.479994 4847 generic.go:334] "Generic (PLEG): container finished" podID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerID="58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27" exitCode=2 Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.479743 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerDied","Data":"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4"} Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.480093 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerDied","Data":"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27"} Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.482736 4847 generic.go:334] "Generic (PLEG): container finished" podID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerID="d3c0fad1dee22d2a9d92e55d0f6fdc48d84ff1c2f9fa712011a28308fed14a33" exitCode=0 Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.482807 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" event={"ID":"159c593e-de1b-4dfe-b846-d0da68a0881a","Type":"ContainerDied","Data":"d3c0fad1dee22d2a9d92e55d0f6fdc48d84ff1c2f9fa712011a28308fed14a33"} Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.505810 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerStarted","Data":"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae"} Dec 10 14:43:44 crc kubenswrapper[4847]: I1210 14:43:44.778491 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" path="/var/lib/kubelet/pods/02e9d21c-37e3-40c4-83dc-c2a00b9df7f8/volumes" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.250420 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.520132 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" event={"ID":"159c593e-de1b-4dfe-b846-d0da68a0881a","Type":"ContainerStarted","Data":"a710b931fedbf0a85ddad2347e83365594a0b23b9583983cc7f44422aaaac81e"} Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.520273 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.523656 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerStarted","Data":"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47"} Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.523841 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api-log" containerID="cri-o://11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" gracePeriod=30 Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.523857 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.523901 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api" containerID="cri-o://4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" gracePeriod=30 Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.540744 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerStarted","Data":"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7"} Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.541100 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerStarted","Data":"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62"} Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.552106 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" podStartSLOduration=6.552086578 podStartE2EDuration="6.552086578s" podCreationTimestamp="2025-12-10 14:43:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:45.544853938 +0000 UTC m=+1175.114071578" watchObservedRunningTime="2025-12-10 14:43:45.552086578 +0000 UTC m=+1175.121304208" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.568893 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.612403824 podStartE2EDuration="6.568869072s" podCreationTimestamp="2025-12-10 14:43:39 +0000 UTC" firstStartedPulling="2025-12-10 14:43:42.875508737 +0000 UTC m=+1172.444726367" lastFinishedPulling="2025-12-10 14:43:43.831973985 +0000 UTC m=+1173.401191615" observedRunningTime="2025-12-10 14:43:45.561880218 +0000 UTC m=+1175.131097848" watchObservedRunningTime="2025-12-10 14:43:45.568869072 +0000 UTC m=+1175.138086702" Dec 10 14:43:45 crc kubenswrapper[4847]: I1210 14:43:45.585379 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.585354627 podStartE2EDuration="6.585354627s" podCreationTimestamp="2025-12-10 14:43:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:45.580480702 +0000 UTC m=+1175.149698332" watchObservedRunningTime="2025-12-10 14:43:45.585354627 +0000 UTC m=+1175.154572257" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.214933 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.224794 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323475 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdnld\" (UniqueName: \"kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323586 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323638 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323692 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323856 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323907 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.323940 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324027 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324063 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324086 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle\") pod \"7470d0c7-1541-4b27-b4b1-5075216514de\" (UID: \"7470d0c7-1541-4b27-b4b1-5075216514de\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324116 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324142 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324176 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgxln\" (UniqueName: \"kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.324216 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data\") pod \"b9ecc8fb-2094-44fd-b157-c654a293036e\" (UID: \"b9ecc8fb-2094-44fd-b157-c654a293036e\") " Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.325828 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.326187 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs" (OuterVolumeSpecName: "logs") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.326429 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.326496 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.331982 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts" (OuterVolumeSpecName: "scripts") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.333093 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld" (OuterVolumeSpecName: "kube-api-access-jdnld") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "kube-api-access-jdnld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.333501 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln" (OuterVolumeSpecName: "kube-api-access-qgxln") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "kube-api-access-qgxln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.335040 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts" (OuterVolumeSpecName: "scripts") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.336076 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.360240 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.370106 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471020 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471052 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471063 4847 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7470d0c7-1541-4b27-b4b1-5075216514de-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471073 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7470d0c7-1541-4b27-b4b1-5075216514de-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471082 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471091 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471099 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471109 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgxln\" (UniqueName: \"kubernetes.io/projected/b9ecc8fb-2094-44fd-b157-c654a293036e-kube-api-access-qgxln\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471118 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdnld\" (UniqueName: \"kubernetes.io/projected/7470d0c7-1541-4b27-b4b1-5075216514de-kube-api-access-jdnld\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471126 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b9ecc8fb-2094-44fd-b157-c654a293036e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.471136 4847 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.473559 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data" (OuterVolumeSpecName: "config-data") pod "7470d0c7-1541-4b27-b4b1-5075216514de" (UID: "7470d0c7-1541-4b27-b4b1-5075216514de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.475352 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.480244 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data" (OuterVolumeSpecName: "config-data") pod "b9ecc8fb-2094-44fd-b157-c654a293036e" (UID: "b9ecc8fb-2094-44fd-b157-c654a293036e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.554118 4847 generic.go:334] "Generic (PLEG): container finished" podID="7470d0c7-1541-4b27-b4b1-5075216514de" containerID="4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" exitCode=0 Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.554801 4847 generic.go:334] "Generic (PLEG): container finished" podID="7470d0c7-1541-4b27-b4b1-5075216514de" containerID="11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" exitCode=143 Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.554997 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerDied","Data":"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47"} Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.555088 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerDied","Data":"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae"} Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.555185 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7470d0c7-1541-4b27-b4b1-5075216514de","Type":"ContainerDied","Data":"5c79f214ba2cee819c9b03c36a943ed310811aa0bbd4ff5047e5ddab1dbd91ef"} Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.555261 4847 scope.go:117] "RemoveContainer" containerID="4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.555433 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.564910 4847 generic.go:334] "Generic (PLEG): container finished" podID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerID="062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3" exitCode=0 Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.566083 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.568687 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerDied","Data":"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3"} Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.568809 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b9ecc8fb-2094-44fd-b157-c654a293036e","Type":"ContainerDied","Data":"169558239df569f96e448d76dd494e20d5f05d2e86385509862f4f534f53b4cb"} Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.572637 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.572670 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7470d0c7-1541-4b27-b4b1-5075216514de-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.572682 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9ecc8fb-2094-44fd-b157-c654a293036e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.584478 4847 scope.go:117] "RemoveContainer" containerID="11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.608170 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.616905 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.655591 4847 scope.go:117] "RemoveContainer" containerID="4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.656007 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47\": container with ID starting with 4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47 not found: ID does not exist" containerID="4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656041 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47"} err="failed to get container status \"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47\": rpc error: code = NotFound desc = could not find container \"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47\": container with ID starting with 4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47 not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656063 4847 scope.go:117] "RemoveContainer" containerID="11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.656258 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae\": container with ID starting with 11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae not found: ID does not exist" containerID="11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656278 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae"} err="failed to get container status \"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae\": rpc error: code = NotFound desc = could not find container \"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae\": container with ID starting with 11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656291 4847 scope.go:117] "RemoveContainer" containerID="4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656447 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47"} err="failed to get container status \"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47\": rpc error: code = NotFound desc = could not find container \"4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47\": container with ID starting with 4aa236422291e4cd487053fcac16b2dcafead49a21dec27f727baba9e3705c47 not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656464 4847 scope.go:117] "RemoveContainer" containerID="11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656956 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae"} err="failed to get container status \"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae\": rpc error: code = NotFound desc = could not find container \"11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae\": container with ID starting with 11d746e8bafb4f88343bf8bdbc3bd982915e4eaa81383b80c5980606fe132fae not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.656979 4847 scope.go:117] "RemoveContainer" containerID="8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.697693 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698106 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="sg-core" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698124 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="sg-core" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698136 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="ceilometer-notification-agent" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698143 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="ceilometer-notification-agent" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698155 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="init" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698160 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="init" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698173 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="dnsmasq-dns" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698180 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="dnsmasq-dns" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698190 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api-log" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698195 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api-log" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698226 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698232 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.698245 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="proxy-httpd" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698252 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="proxy-httpd" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698429 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api-log" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698445 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="ceilometer-notification-agent" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698456 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e9d21c-37e3-40c4-83dc-c2a00b9df7f8" containerName="dnsmasq-dns" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698463 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="sg-core" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698473 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" containerName="proxy-httpd" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.698489 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" containerName="cinder-api" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.699630 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.710261 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.710790 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.710935 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.722127 4847 scope.go:117] "RemoveContainer" containerID="58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.733659 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.747506 4847 scope.go:117] "RemoveContainer" containerID="062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.752595 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778788 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczrn\" (UniqueName: \"kubernetes.io/projected/f0c3090a-89da-46f1-9998-a68188814b05-kube-api-access-hczrn\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778844 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778869 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778927 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data-custom\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778953 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778979 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0c3090a-89da-46f1-9998-a68188814b05-logs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.778996 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0c3090a-89da-46f1-9998-a68188814b05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.779049 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-scripts\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.779079 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.792097 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7470d0c7-1541-4b27-b4b1-5075216514de" path="/var/lib/kubelet/pods/7470d0c7-1541-4b27-b4b1-5075216514de/volumes" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.792991 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9ecc8fb-2094-44fd-b157-c654a293036e" path="/var/lib/kubelet/pods/b9ecc8fb-2094-44fd-b157-c654a293036e/volumes" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.794102 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.794141 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.797924 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.802089 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.802399 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.802532 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.821998 4847 scope.go:117] "RemoveContainer" containerID="8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.822919 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4\": container with ID starting with 8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4 not found: ID does not exist" containerID="8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.822951 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4"} err="failed to get container status \"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4\": rpc error: code = NotFound desc = could not find container \"8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4\": container with ID starting with 8434a182d6020d43ed25fac032c49d7ba5a9b09aaf9eaf0b86d344654051cae4 not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.822976 4847 scope.go:117] "RemoveContainer" containerID="58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.823470 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27\": container with ID starting with 58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27 not found: ID does not exist" containerID="58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.823494 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27"} err="failed to get container status \"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27\": rpc error: code = NotFound desc = could not find container \"58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27\": container with ID starting with 58d6a83d47ea95e70a1079f1c9237f0599865aad61ca55487318376673f56a27 not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.823512 4847 scope.go:117] "RemoveContainer" containerID="062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3" Dec 10 14:43:46 crc kubenswrapper[4847]: E1210 14:43:46.824790 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3\": container with ID starting with 062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3 not found: ID does not exist" containerID="062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.824819 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3"} err="failed to get container status \"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3\": rpc error: code = NotFound desc = could not find container \"062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3\": container with ID starting with 062e1858e2a87d494f53f3166474e5598ef9a24b3feac746a94124547545aaf3 not found: ID does not exist" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.881250 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.881364 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-scripts\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.882165 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.882230 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.882298 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczrn\" (UniqueName: \"kubernetes.io/projected/f0c3090a-89da-46f1-9998-a68188814b05-kube-api-access-hczrn\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.882381 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.882414 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884410 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data-custom\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884454 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg4p6\" (UniqueName: \"kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884518 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884618 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884647 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0c3090a-89da-46f1-9998-a68188814b05-logs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884674 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0c3090a-89da-46f1-9998-a68188814b05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884749 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884778 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.884838 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.888231 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0c3090a-89da-46f1-9998-a68188814b05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.889310 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-scripts\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.906418 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.909879 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0c3090a-89da-46f1-9998-a68188814b05-logs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.910247 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.910850 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.910905 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.917846 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0c3090a-89da-46f1-9998-a68188814b05-config-data-custom\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.921692 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczrn\" (UniqueName: \"kubernetes.io/projected/f0c3090a-89da-46f1-9998-a68188814b05-kube-api-access-hczrn\") pod \"cinder-api-0\" (UID: \"f0c3090a-89da-46f1-9998-a68188814b05\") " pod="openstack/cinder-api-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986403 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg4p6\" (UniqueName: \"kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986484 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986551 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986587 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986613 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.986681 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.988553 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.989208 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.991495 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.992043 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.993732 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:46 crc kubenswrapper[4847]: I1210 14:43:46.998008 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.007032 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg4p6\" (UniqueName: \"kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6\") pod \"ceilometer-0\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " pod="openstack/ceilometer-0" Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.030895 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.148099 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.530204 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 14:43:47 crc kubenswrapper[4847]: W1210 14:43:47.557868 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0c3090a_89da_46f1_9998_a68188814b05.slice/crio-a2acf9083de0e186a54ea68a5fd560d530111f620e91131132f5e863abbe0b8f WatchSource:0}: Error finding container a2acf9083de0e186a54ea68a5fd560d530111f620e91131132f5e863abbe0b8f: Status 404 returned error can't find the container with id a2acf9083de0e186a54ea68a5fd560d530111f620e91131132f5e863abbe0b8f Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.581985 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f0c3090a-89da-46f1-9998-a68188814b05","Type":"ContainerStarted","Data":"a2acf9083de0e186a54ea68a5fd560d530111f620e91131132f5e863abbe0b8f"} Dec 10 14:43:47 crc kubenswrapper[4847]: I1210 14:43:47.664131 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:43:48 crc kubenswrapper[4847]: I1210 14:43:48.601120 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerStarted","Data":"f17376aed3e3b044cbe8a08d1867f4dd1be47d41042e41a14abe38516a4bd620"} Dec 10 14:43:48 crc kubenswrapper[4847]: I1210 14:43:48.609068 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f0c3090a-89da-46f1-9998-a68188814b05","Type":"ContainerStarted","Data":"df3e0449c19f4b57fb14ed8e1dfce4fdc8434edb07c71ff130dc3a50e05454be"} Dec 10 14:43:48 crc kubenswrapper[4847]: I1210 14:43:48.773533 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:43:48 crc kubenswrapper[4847]: I1210 14:43:48.864811 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.074028 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5488ccdfb4-gmxkx" Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.080210 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c86c8879c-drvf5" Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.169686 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.170238 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7587c5f9d-h9fj2" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" containerID="cri-o://1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67" gracePeriod=30 Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.170702 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7587c5f9d-h9fj2" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api" containerID="cri-o://9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421" gracePeriod=30 Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.188989 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.189384 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c4f9cd7f4-7j84b" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-api" containerID="cri-o://7a9278c853165bbb366d2b453ee252212864e89a1975e0e169e6fb53f88b09ca" gracePeriod=30 Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.190018 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c4f9cd7f4-7j84b" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-httpd" containerID="cri-o://354a8d3c17fca29ecffa725ea38a20b3774996503a0f71e9e2ca6373d6267c4e" gracePeriod=30 Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.648243 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f0c3090a-89da-46f1-9998-a68188814b05","Type":"ContainerStarted","Data":"432c2f117dd855d2c670acb3fbebedccb30661858dc249e765fe34a37f1b2573"} Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.650334 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.674919 4847 generic.go:334] "Generic (PLEG): container finished" podID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerID="1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67" exitCode=143 Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.675065 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerDied","Data":"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67"} Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.680322 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerStarted","Data":"6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7"} Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.703092 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.703067631 podStartE2EDuration="3.703067631s" podCreationTimestamp="2025-12-10 14:43:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:49.675412269 +0000 UTC m=+1179.244629889" watchObservedRunningTime="2025-12-10 14:43:49.703067631 +0000 UTC m=+1179.272285261" Dec 10 14:43:49 crc kubenswrapper[4847]: I1210 14:43:49.988830 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.098375 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.205883 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.206303 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="dnsmasq-dns" containerID="cri-o://c940e57e6381a801026597659fc6417f1b1fc8005c80ec459db203abdbc7599d" gracePeriod=10 Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.391585 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.731248 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerID="354a8d3c17fca29ecffa725ea38a20b3774996503a0f71e9e2ca6373d6267c4e" exitCode=0 Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.731662 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerDied","Data":"354a8d3c17fca29ecffa725ea38a20b3774996503a0f71e9e2ca6373d6267c4e"} Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.740724 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerStarted","Data":"67809df9b986ec7f398acc37b38e61c3eb6dc8b23dbf6b0a3ef31a4f37924905"} Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.747163 4847 generic.go:334] "Generic (PLEG): container finished" podID="3152f661-df13-4d77-a736-ddf382434a2c" containerID="c940e57e6381a801026597659fc6417f1b1fc8005c80ec459db203abdbc7599d" exitCode=0 Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.747613 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" event={"ID":"3152f661-df13-4d77-a736-ddf382434a2c","Type":"ContainerDied","Data":"c940e57e6381a801026597659fc6417f1b1fc8005c80ec459db203abdbc7599d"} Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.802983 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.827857 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945063 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945437 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlxxz\" (UniqueName: \"kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945497 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945599 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945631 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.945660 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config\") pod \"3152f661-df13-4d77-a736-ddf382434a2c\" (UID: \"3152f661-df13-4d77-a736-ddf382434a2c\") " Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.951236 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz" (OuterVolumeSpecName: "kube-api-access-jlxxz") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "kube-api-access-jlxxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:50 crc kubenswrapper[4847]: I1210 14:43:50.997461 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.001200 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.003660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config" (OuterVolumeSpecName: "config") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.006796 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.016425 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3152f661-df13-4d77-a736-ddf382434a2c" (UID: "3152f661-df13-4d77-a736-ddf382434a2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048356 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048396 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlxxz\" (UniqueName: \"kubernetes.io/projected/3152f661-df13-4d77-a736-ddf382434a2c-kube-api-access-jlxxz\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048407 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048417 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048428 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.048438 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3152f661-df13-4d77-a736-ddf382434a2c-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.138039 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.626384 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d4f6b7576-wvhs4" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.681136 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.765423 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerStarted","Data":"9361b4cbb56a697b672b2abd36acff003cfdf46044625ddd95979737052e93d2"} Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.770218 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="cinder-scheduler" containerID="cri-o://4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62" gracePeriod=30 Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.770968 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="probe" containerID="cri-o://b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7" gracePeriod=30 Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.771074 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.771789 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-bwcb9" event={"ID":"3152f661-df13-4d77-a736-ddf382434a2c","Type":"ContainerDied","Data":"e3718bec7015c83b6066ae1624bf1a5b542d6a3fc4e5cf6759e6501dc0b0b649"} Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.772027 4847 scope.go:117] "RemoveContainer" containerID="c940e57e6381a801026597659fc6417f1b1fc8005c80ec459db203abdbc7599d" Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.772330 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon-log" containerID="cri-o://d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f" gracePeriod=30 Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.772512 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" containerID="cri-o://54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50" gracePeriod=30 Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.831784 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.842050 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-bwcb9"] Dec 10 14:43:51 crc kubenswrapper[4847]: I1210 14:43:51.844882 4847 scope.go:117] "RemoveContainer" containerID="f60f1ccd2e32d8c810ba0ee6bccd1dbcbda256bd8286ff17a53458239efe52ef" Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.781662 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3152f661-df13-4d77-a736-ddf382434a2c" path="/var/lib/kubelet/pods/3152f661-df13-4d77-a736-ddf382434a2c/volumes" Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.790398 4847 generic.go:334] "Generic (PLEG): container finished" podID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerID="b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7" exitCode=0 Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.790473 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerDied","Data":"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7"} Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.792454 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerStarted","Data":"2706634cb15ea4e769ea87d5048a7742eb3011268cad432a43486cfcba5bd849"} Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.792983 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.820259 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7587c5f9d-h9fj2" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:60902->10.217.0.158:9311: read: connection reset by peer" Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.820535 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7587c5f9d-h9fj2" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:60888->10.217.0.158:9311: read: connection reset by peer" Dec 10 14:43:52 crc kubenswrapper[4847]: I1210 14:43:52.829266 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.034617724 podStartE2EDuration="6.82924199s" podCreationTimestamp="2025-12-10 14:43:46 +0000 UTC" firstStartedPulling="2025-12-10 14:43:47.681445197 +0000 UTC m=+1177.250662827" lastFinishedPulling="2025-12-10 14:43:52.476069463 +0000 UTC m=+1182.045287093" observedRunningTime="2025-12-10 14:43:52.817200858 +0000 UTC m=+1182.386418488" watchObservedRunningTime="2025-12-10 14:43:52.82924199 +0000 UTC m=+1182.398459620" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.247879 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.402786 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs\") pod \"abd57d8c-02f5-4183-921b-a27b1d99eaae\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.402852 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle\") pod \"abd57d8c-02f5-4183-921b-a27b1d99eaae\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.402941 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data\") pod \"abd57d8c-02f5-4183-921b-a27b1d99eaae\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.403045 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom\") pod \"abd57d8c-02f5-4183-921b-a27b1d99eaae\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.403107 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzndd\" (UniqueName: \"kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd\") pod \"abd57d8c-02f5-4183-921b-a27b1d99eaae\" (UID: \"abd57d8c-02f5-4183-921b-a27b1d99eaae\") " Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.403659 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs" (OuterVolumeSpecName: "logs") pod "abd57d8c-02f5-4183-921b-a27b1d99eaae" (UID: "abd57d8c-02f5-4183-921b-a27b1d99eaae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.408955 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "abd57d8c-02f5-4183-921b-a27b1d99eaae" (UID: "abd57d8c-02f5-4183-921b-a27b1d99eaae"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.409177 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd" (OuterVolumeSpecName: "kube-api-access-pzndd") pod "abd57d8c-02f5-4183-921b-a27b1d99eaae" (UID: "abd57d8c-02f5-4183-921b-a27b1d99eaae"). InnerVolumeSpecName "kube-api-access-pzndd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.432271 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abd57d8c-02f5-4183-921b-a27b1d99eaae" (UID: "abd57d8c-02f5-4183-921b-a27b1d99eaae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.463961 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data" (OuterVolumeSpecName: "config-data") pod "abd57d8c-02f5-4183-921b-a27b1d99eaae" (UID: "abd57d8c-02f5-4183-921b-a27b1d99eaae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.505593 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzndd\" (UniqueName: \"kubernetes.io/projected/abd57d8c-02f5-4183-921b-a27b1d99eaae-kube-api-access-pzndd\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.505920 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd57d8c-02f5-4183-921b-a27b1d99eaae-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.506007 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.506110 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.506202 4847 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd57d8c-02f5-4183-921b-a27b1d99eaae-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.806211 4847 generic.go:334] "Generic (PLEG): container finished" podID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerID="9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421" exitCode=0 Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.806248 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerDied","Data":"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421"} Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.807400 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7587c5f9d-h9fj2" event={"ID":"abd57d8c-02f5-4183-921b-a27b1d99eaae","Type":"ContainerDied","Data":"7ad994b83d9fff91f53501bf519bc01d2288f7d2c3cae9c275388ff9ae8cf73c"} Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.807422 4847 scope.go:117] "RemoveContainer" containerID="9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.806310 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7587c5f9d-h9fj2" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.840109 4847 scope.go:117] "RemoveContainer" containerID="1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.848060 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.858923 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7587c5f9d-h9fj2"] Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.870090 4847 scope.go:117] "RemoveContainer" containerID="9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421" Dec 10 14:43:53 crc kubenswrapper[4847]: E1210 14:43:53.870854 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421\": container with ID starting with 9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421 not found: ID does not exist" containerID="9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.870919 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421"} err="failed to get container status \"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421\": rpc error: code = NotFound desc = could not find container \"9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421\": container with ID starting with 9d491a814f500d2438cafd2ea3c16480a8cb3d46f152b7864ad5e83c22140421 not found: ID does not exist" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.870941 4847 scope.go:117] "RemoveContainer" containerID="1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67" Dec 10 14:43:53 crc kubenswrapper[4847]: E1210 14:43:53.871264 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67\": container with ID starting with 1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67 not found: ID does not exist" containerID="1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67" Dec 10 14:43:53 crc kubenswrapper[4847]: I1210 14:43:53.871283 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67"} err="failed to get container status \"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67\": rpc error: code = NotFound desc = could not find container \"1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67\": container with ID starting with 1c4adf6ef5f85a48145d424cef9cbd30e965490d58675f0844101d8e12e4bb67 not found: ID does not exist" Dec 10 14:43:54 crc kubenswrapper[4847]: I1210 14:43:54.772637 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" path="/var/lib/kubelet/pods/abd57d8c-02f5-4183-921b-a27b1d99eaae/volumes" Dec 10 14:43:54 crc kubenswrapper[4847]: I1210 14:43:54.816829 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerID="7a9278c853165bbb366d2b453ee252212864e89a1975e0e169e6fb53f88b09ca" exitCode=0 Dec 10 14:43:54 crc kubenswrapper[4847]: I1210 14:43:54.816904 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerDied","Data":"7a9278c853165bbb366d2b453ee252212864e89a1975e0e169e6fb53f88b09ca"} Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.129702 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.241008 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ps9lk\" (UniqueName: \"kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk\") pod \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.241152 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config\") pod \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.241222 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs\") pod \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.241272 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle\") pod \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.241293 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config\") pod \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\" (UID: \"8a44c113-e0fa-47c6-afa1-909fbe3852b1\") " Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.246791 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8a44c113-e0fa-47c6-afa1-909fbe3852b1" (UID: "8a44c113-e0fa-47c6-afa1-909fbe3852b1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.250744 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk" (OuterVolumeSpecName: "kube-api-access-ps9lk") pod "8a44c113-e0fa-47c6-afa1-909fbe3852b1" (UID: "8a44c113-e0fa-47c6-afa1-909fbe3852b1"). InnerVolumeSpecName "kube-api-access-ps9lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.301821 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config" (OuterVolumeSpecName: "config") pod "8a44c113-e0fa-47c6-afa1-909fbe3852b1" (UID: "8a44c113-e0fa-47c6-afa1-909fbe3852b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.311187 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a44c113-e0fa-47c6-afa1-909fbe3852b1" (UID: "8a44c113-e0fa-47c6-afa1-909fbe3852b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.325403 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8a44c113-e0fa-47c6-afa1-909fbe3852b1" (UID: "8a44c113-e0fa-47c6-afa1-909fbe3852b1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.343346 4847 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.343386 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.343396 4847 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.343405 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ps9lk\" (UniqueName: \"kubernetes.io/projected/8a44c113-e0fa-47c6-afa1-909fbe3852b1-kube-api-access-ps9lk\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.343417 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8a44c113-e0fa-47c6-afa1-909fbe3852b1-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.829484 4847 generic.go:334] "Generic (PLEG): container finished" podID="517d4777-6273-401e-bf27-b23ea11f1a41" containerID="54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50" exitCode=0 Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.829588 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerDied","Data":"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50"} Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.832967 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c4f9cd7f4-7j84b" event={"ID":"8a44c113-e0fa-47c6-afa1-909fbe3852b1","Type":"ContainerDied","Data":"95518f0b60049b75e8f6be5d82481be8b83b8881de27791f4602ba7bd588c02f"} Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.833015 4847 scope.go:117] "RemoveContainer" containerID="354a8d3c17fca29ecffa725ea38a20b3774996503a0f71e9e2ca6373d6267c4e" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.833086 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c4f9cd7f4-7j84b" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.871648 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.873665 4847 scope.go:117] "RemoveContainer" containerID="7a9278c853165bbb366d2b453ee252212864e89a1975e0e169e6fb53f88b09ca" Dec 10 14:43:55 crc kubenswrapper[4847]: I1210 14:43:55.880003 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6c4f9cd7f4-7j84b"] Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.532510 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.683650 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.771037 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" path="/var/lib/kubelet/pods/8a44c113-e0fa-47c6-afa1-909fbe3852b1/volumes" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.773942 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhnvw\" (UniqueName: \"kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.774949 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.775030 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.775098 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.775130 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.775231 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts\") pod \"5f4eb914-5bc6-4e28-9466-4fec9136b183\" (UID: \"5f4eb914-5bc6-4e28-9466-4fec9136b183\") " Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.776025 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.780482 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw" (OuterVolumeSpecName: "kube-api-access-zhnvw") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "kube-api-access-zhnvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.782809 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.783344 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts" (OuterVolumeSpecName: "scripts") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.843992 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.885079 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.885112 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.885125 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhnvw\" (UniqueName: \"kubernetes.io/projected/5f4eb914-5bc6-4e28-9466-4fec9136b183-kube-api-access-zhnvw\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.885136 4847 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5f4eb914-5bc6-4e28-9466-4fec9136b183-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.885146 4847 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.901191 4847 generic.go:334] "Generic (PLEG): container finished" podID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerID="4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62" exitCode=0 Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.901271 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerDied","Data":"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62"} Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.901305 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5f4eb914-5bc6-4e28-9466-4fec9136b183","Type":"ContainerDied","Data":"e9b7afe6e85fc31d8a000795732034193c7b52ebaa895471d0636583ff9957a1"} Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.901344 4847 scope.go:117] "RemoveContainer" containerID="b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.901545 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.948134 4847 scope.go:117] "RemoveContainer" containerID="4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.948491 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data" (OuterVolumeSpecName: "config-data") pod "5f4eb914-5bc6-4e28-9466-4fec9136b183" (UID: "5f4eb914-5bc6-4e28-9466-4fec9136b183"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.977764 4847 scope.go:117] "RemoveContainer" containerID="b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7" Dec 10 14:43:56 crc kubenswrapper[4847]: E1210 14:43:56.978685 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7\": container with ID starting with b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7 not found: ID does not exist" containerID="b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.978849 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7"} err="failed to get container status \"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7\": rpc error: code = NotFound desc = could not find container \"b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7\": container with ID starting with b63a50c1905cebac9b847b43b46dcf012d74a1aba06af9905989805799ec3fd7 not found: ID does not exist" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.978955 4847 scope.go:117] "RemoveContainer" containerID="4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62" Dec 10 14:43:56 crc kubenswrapper[4847]: E1210 14:43:56.979351 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62\": container with ID starting with 4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62 not found: ID does not exist" containerID="4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.979395 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62"} err="failed to get container status \"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62\": rpc error: code = NotFound desc = could not find container \"4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62\": container with ID starting with 4bb350267287c772e83b86cda82d5e53873ac09d2f06e80bf60520abb2eb1f62 not found: ID does not exist" Dec 10 14:43:56 crc kubenswrapper[4847]: I1210 14:43:56.987876 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f4eb914-5bc6-4e28-9466-4fec9136b183-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.242085 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.251188 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.277696 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278139 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278195 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278215 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278227 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278245 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="init" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278256 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="init" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278280 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="dnsmasq-dns" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278289 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="dnsmasq-dns" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278308 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="cinder-scheduler" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278319 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="cinder-scheduler" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278332 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="probe" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278342 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="probe" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278364 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-api" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278373 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-api" Dec 10 14:43:57 crc kubenswrapper[4847]: E1210 14:43:57.278392 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-httpd" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278403 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-httpd" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278612 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api-log" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278635 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="probe" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278649 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd57d8c-02f5-4183-921b-a27b1d99eaae" containerName="barbican-api" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278664 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" containerName="cinder-scheduler" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278680 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-httpd" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278695 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a44c113-e0fa-47c6-afa1-909fbe3852b1" containerName="neutron-api" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.278738 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="3152f661-df13-4d77-a736-ddf382434a2c" containerName="dnsmasq-dns" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.279939 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.287235 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.295985 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402148 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qp6\" (UniqueName: \"kubernetes.io/projected/6b8b833f-7fa3-4232-831a-18d4ba96304b-kube-api-access-65qp6\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402219 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402282 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402330 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402384 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b8b833f-7fa3-4232-831a-18d4ba96304b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.402412 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.468221 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503554 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qp6\" (UniqueName: \"kubernetes.io/projected/6b8b833f-7fa3-4232-831a-18d4ba96304b-kube-api-access-65qp6\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503607 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503643 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503669 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503700 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b8b833f-7fa3-4232-831a-18d4ba96304b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503774 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.503830 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b8b833f-7fa3-4232-831a-18d4ba96304b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.509752 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.515366 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.516290 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.519208 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qp6\" (UniqueName: \"kubernetes.io/projected/6b8b833f-7fa3-4232-831a-18d4ba96304b-kube-api-access-65qp6\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.528954 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8b833f-7fa3-4232-831a-18d4ba96304b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6b8b833f-7fa3-4232-831a-18d4ba96304b\") " pod="openstack/cinder-scheduler-0" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.605501 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6d896bd648-rk9qk" Dec 10 14:43:57 crc kubenswrapper[4847]: I1210 14:43:57.609131 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 14:43:58 crc kubenswrapper[4847]: I1210 14:43:58.157640 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 14:43:58 crc kubenswrapper[4847]: W1210 14:43:58.162357 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b8b833f_7fa3_4232_831a_18d4ba96304b.slice/crio-befd93e0cf679067ffcebe4d27db8c2d8d7d10b938eece78246a469a2cb27b82 WatchSource:0}: Error finding container befd93e0cf679067ffcebe4d27db8c2d8d7d10b938eece78246a469a2cb27b82: Status 404 returned error can't find the container with id befd93e0cf679067ffcebe4d27db8c2d8d7d10b938eece78246a469a2cb27b82 Dec 10 14:43:58 crc kubenswrapper[4847]: I1210 14:43:58.780461 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4eb914-5bc6-4e28-9466-4fec9136b183" path="/var/lib/kubelet/pods/5f4eb914-5bc6-4e28-9466-4fec9136b183/volumes" Dec 10 14:43:58 crc kubenswrapper[4847]: I1210 14:43:58.924447 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b8b833f-7fa3-4232-831a-18d4ba96304b","Type":"ContainerStarted","Data":"1651dea3513f8ac59a00ac999353b11493de26aa1ec0b2de2e57baee8a465fdb"} Dec 10 14:43:58 crc kubenswrapper[4847]: I1210 14:43:58.924489 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b8b833f-7fa3-4232-831a-18d4ba96304b","Type":"ContainerStarted","Data":"befd93e0cf679067ffcebe4d27db8c2d8d7d10b938eece78246a469a2cb27b82"} Dec 10 14:43:59 crc kubenswrapper[4847]: I1210 14:43:59.315089 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 14:43:59 crc kubenswrapper[4847]: I1210 14:43:59.935208 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6b8b833f-7fa3-4232-831a-18d4ba96304b","Type":"ContainerStarted","Data":"f1e803c8229bdb57820ada7f070ff4de308af3aa604b15767e599020247cb7a2"} Dec 10 14:43:59 crc kubenswrapper[4847]: I1210 14:43:59.954268 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.954254385 podStartE2EDuration="2.954254385s" podCreationTimestamp="2025-12-10 14:43:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:43:59.95300814 +0000 UTC m=+1189.522225760" watchObservedRunningTime="2025-12-10 14:43:59.954254385 +0000 UTC m=+1189.523472015" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.077690 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-54599d665c-kl29l" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.790104 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.791535 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.793991 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.801125 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.801432 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ms5p7" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.806190 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.884506 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.884566 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.884652 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkvh\" (UniqueName: \"kubernetes.io/projected/2559a413-0d27-43f0-8fa3-2da6e9d66eff-kube-api-access-9qkvh\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.884697 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config-secret\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.986453 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.986512 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.986575 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkvh\" (UniqueName: \"kubernetes.io/projected/2559a413-0d27-43f0-8fa3-2da6e9d66eff-kube-api-access-9qkvh\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.986605 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config-secret\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.987858 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.993465 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:01 crc kubenswrapper[4847]: I1210 14:44:01.996182 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2559a413-0d27-43f0-8fa3-2da6e9d66eff-openstack-config-secret\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:02 crc kubenswrapper[4847]: I1210 14:44:02.017583 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkvh\" (UniqueName: \"kubernetes.io/projected/2559a413-0d27-43f0-8fa3-2da6e9d66eff-kube-api-access-9qkvh\") pod \"openstackclient\" (UID: \"2559a413-0d27-43f0-8fa3-2da6e9d66eff\") " pod="openstack/openstackclient" Dec 10 14:44:02 crc kubenswrapper[4847]: I1210 14:44:02.117142 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 14:44:02 crc kubenswrapper[4847]: I1210 14:44:02.589880 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 14:44:02 crc kubenswrapper[4847]: I1210 14:44:02.609350 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 14:44:03 crc kubenswrapper[4847]: I1210 14:44:03.015885 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2559a413-0d27-43f0-8fa3-2da6e9d66eff","Type":"ContainerStarted","Data":"a6521a9f3a76391bdf6d9a27c02f3d147adab2caa3c93ff04ea49dc163821b0c"} Dec 10 14:44:06 crc kubenswrapper[4847]: I1210 14:44:06.531853 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.527056 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65d99c95b7-t6kg4"] Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.529197 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.532298 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.533089 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.533275 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.565578 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65d99c95b7-t6kg4"] Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702436 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-config-data\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702562 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-run-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702655 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-public-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702755 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-etc-swift\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702797 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-internal-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702827 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-combined-ca-bundle\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702853 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf78r\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-kube-api-access-cf78r\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.702892 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-log-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804746 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-public-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804835 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-etc-swift\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804862 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-internal-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804885 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-combined-ca-bundle\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804911 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf78r\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-kube-api-access-cf78r\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804929 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-log-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.804967 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-config-data\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.805018 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-run-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.805626 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-run-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.805688 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cd88676-59ca-4c55-aaf7-caab8965d2bb-log-httpd\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.811048 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-internal-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.811740 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-etc-swift\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.814760 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-public-tls-certs\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.823577 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-config-data\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.826923 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf78r\" (UniqueName: \"kubernetes.io/projected/9cd88676-59ca-4c55-aaf7-caab8965d2bb-kube-api-access-cf78r\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.833194 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cd88676-59ca-4c55-aaf7-caab8965d2bb-combined-ca-bundle\") pod \"swift-proxy-65d99c95b7-t6kg4\" (UID: \"9cd88676-59ca-4c55-aaf7-caab8965d2bb\") " pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.842458 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 14:44:07 crc kubenswrapper[4847]: I1210 14:44:07.865583 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.205255 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.206970 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-central-agent" containerID="cri-o://6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7" gracePeriod=30 Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.207155 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="sg-core" containerID="cri-o://9361b4cbb56a697b672b2abd36acff003cfdf46044625ddd95979737052e93d2" gracePeriod=30 Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.207218 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="proxy-httpd" containerID="cri-o://2706634cb15ea4e769ea87d5048a7742eb3011268cad432a43486cfcba5bd849" gracePeriod=30 Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.207276 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-notification-agent" containerID="cri-o://67809df9b986ec7f398acc37b38e61c3eb6dc8b23dbf6b0a3ef31a4f37924905" gracePeriod=30 Dec 10 14:44:08 crc kubenswrapper[4847]: I1210 14:44:08.216756 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": EOF" Dec 10 14:44:08 crc kubenswrapper[4847]: E1210 14:44:08.750145 4847 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c9841e_2a1b_4796_99ca_cd6eeac1b5bc.slice/crio-conmon-6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7.scope\": RecentStats: unable to find data in memory cache]" Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075180 4847 generic.go:334] "Generic (PLEG): container finished" podID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerID="2706634cb15ea4e769ea87d5048a7742eb3011268cad432a43486cfcba5bd849" exitCode=0 Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075208 4847 generic.go:334] "Generic (PLEG): container finished" podID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerID="9361b4cbb56a697b672b2abd36acff003cfdf46044625ddd95979737052e93d2" exitCode=2 Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075217 4847 generic.go:334] "Generic (PLEG): container finished" podID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerID="6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7" exitCode=0 Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075239 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerDied","Data":"2706634cb15ea4e769ea87d5048a7742eb3011268cad432a43486cfcba5bd849"} Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075263 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerDied","Data":"9361b4cbb56a697b672b2abd36acff003cfdf46044625ddd95979737052e93d2"} Dec 10 14:44:09 crc kubenswrapper[4847]: I1210 14:44:09.075273 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerDied","Data":"6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7"} Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.175301 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-b8f8x"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.176864 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.203099 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-b8f8x"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.275384 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-fsqtk"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.276794 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.293257 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-64b8-account-create-update-6q5jl"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.294698 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.298279 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.309826 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fsqtk"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.322177 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-64b8-account-create-update-6q5jl"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.353966 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.354391 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjqqr\" (UniqueName: \"kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.378690 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bkdzb"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.379891 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.389937 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bkdzb"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456737 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456789 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjqqr\" (UniqueName: \"kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456816 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456844 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7qcj\" (UniqueName: \"kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.456967 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gtfv\" (UniqueName: \"kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.457814 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.485395 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-d0de-account-create-update-k2qzx"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.487379 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.508340 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d0de-account-create-update-k2qzx"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.521325 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.523246 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjqqr\" (UniqueName: \"kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr\") pod \"nova-api-db-create-b8f8x\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.558813 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.558871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.558898 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7qcj\" (UniqueName: \"kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.558975 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2gq\" (UniqueName: \"kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.559006 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gtfv\" (UniqueName: \"kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.559022 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.559615 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.560262 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.575341 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7qcj\" (UniqueName: \"kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj\") pod \"nova-cell0-db-create-fsqtk\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.575364 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gtfv\" (UniqueName: \"kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv\") pod \"nova-api-64b8-account-create-update-6q5jl\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.600628 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.629578 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.661213 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snx2n\" (UniqueName: \"kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.661321 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.661383 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2gq\" (UniqueName: \"kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.661447 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.662230 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.677055 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9a76-account-create-update-6lfrc"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.678473 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.681165 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.693921 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9a76-account-create-update-6lfrc"] Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.715074 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2gq\" (UniqueName: \"kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq\") pod \"nova-cell1-db-create-bkdzb\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.721358 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.765375 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxb5k\" (UniqueName: \"kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.765496 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.765706 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.765749 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snx2n\" (UniqueName: \"kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.766558 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.790771 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snx2n\" (UniqueName: \"kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n\") pod \"nova-cell0-d0de-account-create-update-k2qzx\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.800189 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.810959 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.867254 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxb5k\" (UniqueName: \"kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.867545 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.868316 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:10 crc kubenswrapper[4847]: I1210 14:44:10.884426 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxb5k\" (UniqueName: \"kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k\") pod \"nova-cell1-9a76-account-create-update-6lfrc\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:11 crc kubenswrapper[4847]: I1210 14:44:11.035288 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:12 crc kubenswrapper[4847]: I1210 14:44:12.106099 4847 generic.go:334] "Generic (PLEG): container finished" podID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerID="67809df9b986ec7f398acc37b38e61c3eb6dc8b23dbf6b0a3ef31a4f37924905" exitCode=0 Dec 10 14:44:12 crc kubenswrapper[4847]: I1210 14:44:12.106180 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerDied","Data":"67809df9b986ec7f398acc37b38e61c3eb6dc8b23dbf6b0a3ef31a4f37924905"} Dec 10 14:44:13 crc kubenswrapper[4847]: I1210 14:44:13.552328 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:13 crc kubenswrapper[4847]: I1210 14:44:13.552895 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-log" containerID="cri-o://2639bd1f20c0ca05880a97984295b32c9c4bc7f6247346082a23c53bd2718b60" gracePeriod=30 Dec 10 14:44:13 crc kubenswrapper[4847]: I1210 14:44:13.553416 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-httpd" containerID="cri-o://109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729" gracePeriod=30 Dec 10 14:44:13 crc kubenswrapper[4847]: I1210 14:44:13.993791 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.138756 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.138908 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg4p6\" (UniqueName: \"kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.138946 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.138981 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.139060 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.139102 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.139119 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml\") pod \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\" (UID: \"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc\") " Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.140261 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.142702 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.151804 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6" (OuterVolumeSpecName: "kube-api-access-qg4p6") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "kube-api-access-qg4p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.153513 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c9841e-2a1b-4796-99ca-cd6eeac1b5bc","Type":"ContainerDied","Data":"f17376aed3e3b044cbe8a08d1867f4dd1be47d41042e41a14abe38516a4bd620"} Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.153558 4847 scope.go:117] "RemoveContainer" containerID="2706634cb15ea4e769ea87d5048a7742eb3011268cad432a43486cfcba5bd849" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.153680 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.158577 4847 generic.go:334] "Generic (PLEG): container finished" podID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerID="2639bd1f20c0ca05880a97984295b32c9c4bc7f6247346082a23c53bd2718b60" exitCode=143 Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.158792 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerDied","Data":"2639bd1f20c0ca05880a97984295b32c9c4bc7f6247346082a23c53bd2718b60"} Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.161582 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts" (OuterVolumeSpecName: "scripts") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.168042 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2559a413-0d27-43f0-8fa3-2da6e9d66eff","Type":"ContainerStarted","Data":"d2ed1c18dc3c38029d73664d94f805826bb331d22b4d7559d4822cad31eb72c0"} Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.199083 4847 scope.go:117] "RemoveContainer" containerID="9361b4cbb56a697b672b2abd36acff003cfdf46044625ddd95979737052e93d2" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.204263 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.223525 4847 scope.go:117] "RemoveContainer" containerID="67809df9b986ec7f398acc37b38e61c3eb6dc8b23dbf6b0a3ef31a4f37924905" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.242986 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg4p6\" (UniqueName: \"kubernetes.io/projected/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-kube-api-access-qg4p6\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.243025 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.243036 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.243047 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.243061 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.246538 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.229834148 podStartE2EDuration="13.246520739s" podCreationTimestamp="2025-12-10 14:44:01 +0000 UTC" firstStartedPulling="2025-12-10 14:44:02.595286535 +0000 UTC m=+1192.164504165" lastFinishedPulling="2025-12-10 14:44:13.611973126 +0000 UTC m=+1203.181190756" observedRunningTime="2025-12-10 14:44:14.193437434 +0000 UTC m=+1203.762655074" watchObservedRunningTime="2025-12-10 14:44:14.246520739 +0000 UTC m=+1203.815738369" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.247308 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65d99c95b7-t6kg4"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.252137 4847 scope.go:117] "RemoveContainer" containerID="6389f957d2e6c99fab144f4f8674d1f75a2e5cedf61b0413f119f8093422a8f7" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.302136 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.324988 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data" (OuterVolumeSpecName: "config-data") pod "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" (UID: "74c9841e-2a1b-4796-99ca-cd6eeac1b5bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.345382 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.345427 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.422124 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-fsqtk"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.430339 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-64b8-account-create-update-6q5jl"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.433541 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.650173 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.659772 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.670933 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:14 crc kubenswrapper[4847]: E1210 14:44:14.671417 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-notification-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671452 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-notification-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: E1210 14:44:14.671468 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-central-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671476 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-central-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: E1210 14:44:14.671489 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="proxy-httpd" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671496 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="proxy-httpd" Dec 10 14:44:14 crc kubenswrapper[4847]: E1210 14:44:14.671519 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="sg-core" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671525 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="sg-core" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671757 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="proxy-httpd" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671776 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-notification-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671785 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="sg-core" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.671794 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" containerName="ceilometer-central-agent" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.673692 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.676422 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.677932 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.702585 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9a76-account-create-update-6lfrc"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704054 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704125 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704159 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704205 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704280 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzflw\" (UniqueName: \"kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704317 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704354 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.704925 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.739199 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bkdzb"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.753840 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.758024 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-b8f8x"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.775982 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c9841e-2a1b-4796-99ca-cd6eeac1b5bc" path="/var/lib/kubelet/pods/74c9841e-2a1b-4796-99ca-cd6eeac1b5bc/volumes" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.776961 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.778135 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d0de-account-create-update-k2qzx"] Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.816746 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.819817 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzflw\" (UniqueName: \"kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.820124 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.820201 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.820363 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.820595 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.820662 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.823600 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.824501 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.825639 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.834438 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.834797 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.835554 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:14 crc kubenswrapper[4847]: I1210 14:44:14.856080 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzflw\" (UniqueName: \"kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw\") pod \"ceilometer-0\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " pod="openstack/ceilometer-0" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.144751 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.199615 4847 generic.go:334] "Generic (PLEG): container finished" podID="b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" containerID="27a23ddfa0d2cd6f6008fe3b9cd1f054779cb51aba9c84bab00027dfdf2bba16" exitCode=0 Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.199844 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fsqtk" event={"ID":"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992","Type":"ContainerDied","Data":"27a23ddfa0d2cd6f6008fe3b9cd1f054779cb51aba9c84bab00027dfdf2bba16"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.199907 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fsqtk" event={"ID":"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992","Type":"ContainerStarted","Data":"b6504ed4cb23e8078298e91c4f3185e8618603e467d126317ad7cbb12b913018"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.230023 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-64b8-account-create-update-6q5jl" event={"ID":"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6","Type":"ContainerStarted","Data":"b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.230080 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-64b8-account-create-update-6q5jl" event={"ID":"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6","Type":"ContainerStarted","Data":"633c098b60d4c39d9bb6f9620747e4825467162a1ea6a013cb4b8d5961e928ad"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.246040 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" event={"ID":"7b6a1839-419a-4f80-9c45-ff37701bb916","Type":"ContainerStarted","Data":"f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.246478 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" event={"ID":"7b6a1839-419a-4f80-9c45-ff37701bb916","Type":"ContainerStarted","Data":"286040590bf3f27ab9fe7d49b541dc9b72920dad78bf336a96c4f782df5e40fd"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.272803 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-64b8-account-create-update-6q5jl" podStartSLOduration=5.272777043 podStartE2EDuration="5.272777043s" podCreationTimestamp="2025-12-10 14:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.249971663 +0000 UTC m=+1204.819189283" watchObservedRunningTime="2025-12-10 14:44:15.272777043 +0000 UTC m=+1204.841994673" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.272848 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkdzb" event={"ID":"193125a6-9a88-47b0-b462-daab3aa61691","Type":"ContainerStarted","Data":"b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.274489 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkdzb" event={"ID":"193125a6-9a88-47b0-b462-daab3aa61691","Type":"ContainerStarted","Data":"fe5d544dd07aac194081afedcd7f0b08bfc9a6e9d21166a51958e34d8d19f3f3"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.283937 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8f8x" event={"ID":"7a4fd885-5fc1-4bc8-9b47-57e3103e9197","Type":"ContainerStarted","Data":"cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.283978 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8f8x" event={"ID":"7a4fd885-5fc1-4bc8-9b47-57e3103e9197","Type":"ContainerStarted","Data":"6b5c99cb531ad15c864571ce331a9519e09fb3bbc320ceeb094aeb635d65bbf9"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.288722 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" podStartSLOduration=5.288695542 podStartE2EDuration="5.288695542s" podCreationTimestamp="2025-12-10 14:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.274592373 +0000 UTC m=+1204.843810003" watchObservedRunningTime="2025-12-10 14:44:15.288695542 +0000 UTC m=+1204.857913172" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.300707 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" event={"ID":"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f","Type":"ContainerStarted","Data":"2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.300781 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" event={"ID":"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f","Type":"ContainerStarted","Data":"8d988c377893d15ecf5fc86c661df1e9bf5a965015d258e5181282d1ec46c53f"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.306896 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65d99c95b7-t6kg4" event={"ID":"9cd88676-59ca-4c55-aaf7-caab8965d2bb","Type":"ContainerStarted","Data":"0fd77a3464b8fcbf6a21e4e5db1f4766ce6297e0202aaedb7019fa89184398c3"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.306930 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65d99c95b7-t6kg4" event={"ID":"9cd88676-59ca-4c55-aaf7-caab8965d2bb","Type":"ContainerStarted","Data":"abcf4aaa178616d3d6e837ab5bf2c134dd95cfade2466b91aa4b739246044302"} Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.306981 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.307080 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.334294 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-bkdzb" podStartSLOduration=5.334243819 podStartE2EDuration="5.334243819s" podCreationTimestamp="2025-12-10 14:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.293109734 +0000 UTC m=+1204.862327364" watchObservedRunningTime="2025-12-10 14:44:15.334243819 +0000 UTC m=+1204.903461439" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.350930 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-b8f8x" podStartSLOduration=5.350906719 podStartE2EDuration="5.350906719s" podCreationTimestamp="2025-12-10 14:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.314061942 +0000 UTC m=+1204.883279582" watchObservedRunningTime="2025-12-10 14:44:15.350906719 +0000 UTC m=+1204.920124349" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.371386 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" podStartSLOduration=5.371369113 podStartE2EDuration="5.371369113s" podCreationTimestamp="2025-12-10 14:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.33499608 +0000 UTC m=+1204.904213720" watchObservedRunningTime="2025-12-10 14:44:15.371369113 +0000 UTC m=+1204.940586743" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.388201 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65d99c95b7-t6kg4" podStartSLOduration=8.388176927 podStartE2EDuration="8.388176927s" podCreationTimestamp="2025-12-10 14:44:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:15.361751408 +0000 UTC m=+1204.930969048" watchObservedRunningTime="2025-12-10 14:44:15.388176927 +0000 UTC m=+1204.957394557" Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.516071 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.516292 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-log" containerID="cri-o://b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640" gracePeriod=30 Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.516544 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-httpd" containerID="cri-o://e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1" gracePeriod=30 Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.775588 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:15 crc kubenswrapper[4847]: I1210 14:44:15.815480 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.317672 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerStarted","Data":"b982a13b755b01c19f6f1a4da8dfee6513461a85a746d96bd585a49e7969f753"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.322633 4847 generic.go:334] "Generic (PLEG): container finished" podID="333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" containerID="2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e" exitCode=0 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.322784 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" event={"ID":"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f","Type":"ContainerDied","Data":"2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.326645 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65d99c95b7-t6kg4" event={"ID":"9cd88676-59ca-4c55-aaf7-caab8965d2bb","Type":"ContainerStarted","Data":"e22d7d5df91b9551db228a0ff746f8ed5010cecc56c983d43cd3d3c4c12fe89a"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.328211 4847 generic.go:334] "Generic (PLEG): container finished" podID="20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" containerID="b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda" exitCode=0 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.328264 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-64b8-account-create-update-6q5jl" event={"ID":"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6","Type":"ContainerDied","Data":"b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.332287 4847 generic.go:334] "Generic (PLEG): container finished" podID="7b6a1839-419a-4f80-9c45-ff37701bb916" containerID="f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b" exitCode=0 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.332359 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" event={"ID":"7b6a1839-419a-4f80-9c45-ff37701bb916","Type":"ContainerDied","Data":"f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.333708 4847 generic.go:334] "Generic (PLEG): container finished" podID="193125a6-9a88-47b0-b462-daab3aa61691" containerID="b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68" exitCode=0 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.333815 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkdzb" event={"ID":"193125a6-9a88-47b0-b462-daab3aa61691","Type":"ContainerDied","Data":"b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.341269 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerID="b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640" exitCode=143 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.341385 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerDied","Data":"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.345519 4847 generic.go:334] "Generic (PLEG): container finished" podID="7a4fd885-5fc1-4bc8-9b47-57e3103e9197" containerID="cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2" exitCode=0 Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.345584 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8f8x" event={"ID":"7a4fd885-5fc1-4bc8-9b47-57e3103e9197","Type":"ContainerDied","Data":"cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2"} Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.532497 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-8676865fc8-rvrhh" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.532640 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.846621 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.904310 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts\") pod \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.904423 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7qcj\" (UniqueName: \"kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj\") pod \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\" (UID: \"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992\") " Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.906105 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" (UID: "b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:16 crc kubenswrapper[4847]: I1210 14:44:16.914446 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj" (OuterVolumeSpecName: "kube-api-access-w7qcj") pod "b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" (UID: "b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992"). InnerVolumeSpecName "kube-api-access-w7qcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.009160 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.009889 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7qcj\" (UniqueName: \"kubernetes.io/projected/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992-kube-api-access-w7qcj\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.355640 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-fsqtk" event={"ID":"b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992","Type":"ContainerDied","Data":"b6504ed4cb23e8078298e91c4f3185e8618603e467d126317ad7cbb12b913018"} Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.355688 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6504ed4cb23e8078298e91c4f3185e8618603e467d126317ad7cbb12b913018" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.355742 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-fsqtk" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.358932 4847 generic.go:334] "Generic (PLEG): container finished" podID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerID="109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729" exitCode=0 Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.359054 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerDied","Data":"109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729"} Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.791278 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.937537 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts\") pod \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.937616 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gtfv\" (UniqueName: \"kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv\") pod \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\" (UID: \"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6\") " Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.939884 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" (UID: "20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:17 crc kubenswrapper[4847]: I1210 14:44:17.943609 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv" (OuterVolumeSpecName: "kube-api-access-2gtfv") pod "20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" (UID: "20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6"). InnerVolumeSpecName "kube-api-access-2gtfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.023477 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.039397 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.040743 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.040767 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gtfv\" (UniqueName: \"kubernetes.io/projected/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6-kube-api-access-2gtfv\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.101882 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.109832 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.143933 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts\") pod \"193125a6-9a88-47b0-b462-daab3aa61691\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.144619 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "193125a6-9a88-47b0-b462-daab3aa61691" (UID: "193125a6-9a88-47b0-b462-daab3aa61691"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.144703 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts\") pod \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.144830 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx2gq\" (UniqueName: \"kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq\") pod \"193125a6-9a88-47b0-b462-daab3aa61691\" (UID: \"193125a6-9a88-47b0-b462-daab3aa61691\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.144851 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxb5k\" (UniqueName: \"kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k\") pod \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\" (UID: \"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.145311 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/193125a6-9a88-47b0-b462-daab3aa61691-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.146049 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" (UID: "333a1d69-c6b6-434e-b0e3-7c43c39a2e1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.149138 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k" (OuterVolumeSpecName: "kube-api-access-bxb5k") pod "333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" (UID: "333a1d69-c6b6-434e-b0e3-7c43c39a2e1f"). InnerVolumeSpecName "kube-api-access-bxb5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.151431 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq" (OuterVolumeSpecName: "kube-api-access-nx2gq") pod "193125a6-9a88-47b0-b462-daab3aa61691" (UID: "193125a6-9a88-47b0-b462-daab3aa61691"). InnerVolumeSpecName "kube-api-access-nx2gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.247021 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts\") pod \"7b6a1839-419a-4f80-9c45-ff37701bb916\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.247093 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjqqr\" (UniqueName: \"kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr\") pod \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.247117 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snx2n\" (UniqueName: \"kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n\") pod \"7b6a1839-419a-4f80-9c45-ff37701bb916\" (UID: \"7b6a1839-419a-4f80-9c45-ff37701bb916\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.247135 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts\") pod \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\" (UID: \"7a4fd885-5fc1-4bc8-9b47-57e3103e9197\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.247511 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b6a1839-419a-4f80-9c45-ff37701bb916" (UID: "7b6a1839-419a-4f80-9c45-ff37701bb916"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.248080 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7a4fd885-5fc1-4bc8-9b47-57e3103e9197" (UID: "7a4fd885-5fc1-4bc8-9b47-57e3103e9197"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.248162 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.248185 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx2gq\" (UniqueName: \"kubernetes.io/projected/193125a6-9a88-47b0-b462-daab3aa61691-kube-api-access-nx2gq\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.248198 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxb5k\" (UniqueName: \"kubernetes.io/projected/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f-kube-api-access-bxb5k\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.248209 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b6a1839-419a-4f80-9c45-ff37701bb916-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.251183 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr" (OuterVolumeSpecName: "kube-api-access-cjqqr") pod "7a4fd885-5fc1-4bc8-9b47-57e3103e9197" (UID: "7a4fd885-5fc1-4bc8-9b47-57e3103e9197"). InnerVolumeSpecName "kube-api-access-cjqqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.251608 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n" (OuterVolumeSpecName: "kube-api-access-snx2n") pod "7b6a1839-419a-4f80-9c45-ff37701bb916" (UID: "7b6a1839-419a-4f80-9c45-ff37701bb916"). InnerVolumeSpecName "kube-api-access-snx2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.349440 4847 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.349774 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjqqr\" (UniqueName: \"kubernetes.io/projected/7a4fd885-5fc1-4bc8-9b47-57e3103e9197-kube-api-access-cjqqr\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.349863 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snx2n\" (UniqueName: \"kubernetes.io/projected/7b6a1839-419a-4f80-9c45-ff37701bb916-kube-api-access-snx2n\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.368631 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkdzb" event={"ID":"193125a6-9a88-47b0-b462-daab3aa61691","Type":"ContainerDied","Data":"fe5d544dd07aac194081afedcd7f0b08bfc9a6e9d21166a51958e34d8d19f3f3"} Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.368703 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe5d544dd07aac194081afedcd7f0b08bfc9a6e9d21166a51958e34d8d19f3f3" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.368724 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkdzb" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.370244 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8f8x" event={"ID":"7a4fd885-5fc1-4bc8-9b47-57e3103e9197","Type":"ContainerDied","Data":"6b5c99cb531ad15c864571ce331a9519e09fb3bbc320ceeb094aeb635d65bbf9"} Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.370342 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b5c99cb531ad15c864571ce331a9519e09fb3bbc320ceeb094aeb635d65bbf9" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.370251 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8f8x" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.371559 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" event={"ID":"333a1d69-c6b6-434e-b0e3-7c43c39a2e1f","Type":"ContainerDied","Data":"8d988c377893d15ecf5fc86c661df1e9bf5a965015d258e5181282d1ec46c53f"} Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.371607 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d988c377893d15ecf5fc86c661df1e9bf5a965015d258e5181282d1ec46c53f" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.371653 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9a76-account-create-update-6lfrc" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.381845 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-64b8-account-create-update-6q5jl" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.381869 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-64b8-account-create-update-6q5jl" event={"ID":"20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6","Type":"ContainerDied","Data":"633c098b60d4c39d9bb6f9620747e4825467162a1ea6a013cb4b8d5961e928ad"} Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.381894 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="633c098b60d4c39d9bb6f9620747e4825467162a1ea6a013cb4b8d5961e928ad" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.383067 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" event={"ID":"7b6a1839-419a-4f80-9c45-ff37701bb916","Type":"ContainerDied","Data":"286040590bf3f27ab9fe7d49b541dc9b72920dad78bf336a96c4f782df5e40fd"} Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.383094 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="286040590bf3f27ab9fe7d49b541dc9b72920dad78bf336a96c4f782df5e40fd" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.383135 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d0de-account-create-update-k2qzx" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.850026 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.968561 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.968638 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdmwr\" (UniqueName: \"kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.968795 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.968832 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.968909 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969027 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969051 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969077 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data\") pod \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\" (UID: \"accad8cb-ba38-44f0-bbac-20c879e8a2b9\") " Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969105 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969188 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs" (OuterVolumeSpecName: "logs") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969651 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.969667 4847 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/accad8cb-ba38-44f0-bbac-20c879e8a2b9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.981587 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr" (OuterVolumeSpecName: "kube-api-access-kdmwr") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "kube-api-access-kdmwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.986218 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts" (OuterVolumeSpecName: "scripts") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:18 crc kubenswrapper[4847]: I1210 14:44:18.986608 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.017100 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.071453 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.071501 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.071514 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdmwr\" (UniqueName: \"kubernetes.io/projected/accad8cb-ba38-44f0-bbac-20c879e8a2b9-kube-api-access-kdmwr\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.071532 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.090362 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.094891 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data" (OuterVolumeSpecName: "config-data") pod "accad8cb-ba38-44f0-bbac-20c879e8a2b9" (UID: "accad8cb-ba38-44f0-bbac-20c879e8a2b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.096869 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.159437 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.172926 4847 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.172963 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/accad8cb-ba38-44f0-bbac-20c879e8a2b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.172974 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274401 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274464 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274504 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274539 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274599 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274643 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjnqj\" (UniqueName: \"kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274845 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.274977 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\" (UID: \"6b67bb20-a54d-4e08-8d05-c68fd77bd38f\") " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.275207 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.275280 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs" (OuterVolumeSpecName: "logs") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.275839 4847 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.275861 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.279538 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.281285 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts" (OuterVolumeSpecName: "scripts") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.284974 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj" (OuterVolumeSpecName: "kube-api-access-hjnqj") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "kube-api-access-hjnqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.319839 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.357389 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.369859 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data" (OuterVolumeSpecName: "config-data") pod "6b67bb20-a54d-4e08-8d05-c68fd77bd38f" (UID: "6b67bb20-a54d-4e08-8d05-c68fd77bd38f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.386906 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.386963 4847 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.386977 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.386987 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.386995 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjnqj\" (UniqueName: \"kubernetes.io/projected/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-kube-api-access-hjnqj\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.387004 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b67bb20-a54d-4e08-8d05-c68fd77bd38f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.398611 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerID="e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1" exitCode=0 Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.398762 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.399028 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerDied","Data":"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1"} Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.399180 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b67bb20-a54d-4e08-8d05-c68fd77bd38f","Type":"ContainerDied","Data":"5e55074c6e6d6f862dbb3be9c6bc04a1b178630bb58a096302288e4503ca500c"} Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.399262 4847 scope.go:117] "RemoveContainer" containerID="e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.405804 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.405806 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"accad8cb-ba38-44f0-bbac-20c879e8a2b9","Type":"ContainerDied","Data":"0d57fec2d5f50462c6ffc403b82c3aa01781dc611e6524253b0fe8a999f991e9"} Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.411873 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerStarted","Data":"fa6064489107e11585425f3a26327327c3ec986550ca11e1d165d6f6ba9eb6bd"} Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.412398 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerStarted","Data":"99e9bbf8f3cf2525103e5a294714dcef51853c25b76f10c5e874dfcb410b6741"} Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.419142 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.442946 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.456834 4847 scope.go:117] "RemoveContainer" containerID="b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.507048 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.512096 4847 scope.go:117] "RemoveContainer" containerID="e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.512547 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1\": container with ID starting with e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1 not found: ID does not exist" containerID="e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.512590 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1"} err="failed to get container status \"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1\": rpc error: code = NotFound desc = could not find container \"e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1\": container with ID starting with e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1 not found: ID does not exist" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.512626 4847 scope.go:117] "RemoveContainer" containerID="b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.512886 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640\": container with ID starting with b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640 not found: ID does not exist" containerID="b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.512945 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640"} err="failed to get container status \"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640\": rpc error: code = NotFound desc = could not find container \"b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640\": container with ID starting with b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640 not found: ID does not exist" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.512966 4847 scope.go:117] "RemoveContainer" containerID="109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.530257 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.551688 4847 scope.go:117] "RemoveContainer" containerID="2639bd1f20c0ca05880a97984295b32c9c4bc7f6247346082a23c53bd2718b60" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.556881 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.576625 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.591216 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592192 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592219 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592237 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193125a6-9a88-47b0-b462-daab3aa61691" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592247 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="193125a6-9a88-47b0-b462-daab3aa61691" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592261 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592270 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592296 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a4fd885-5fc1-4bc8-9b47-57e3103e9197" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592306 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a4fd885-5fc1-4bc8-9b47-57e3103e9197" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592325 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592333 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592345 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592352 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592367 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592374 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592390 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592398 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592408 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592416 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: E1210 14:44:19.592427 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6a1839-419a-4f80-9c45-ff37701bb916" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592434 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6a1839-419a-4f80-9c45-ff37701bb916" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592597 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592610 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="193125a6-9a88-47b0-b462-daab3aa61691" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592621 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592627 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592636 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592643 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" containerName="glance-log" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592651 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592661 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" containerName="glance-httpd" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592671 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6a1839-419a-4f80-9c45-ff37701bb916" containerName="mariadb-account-create-update" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.592684 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a4fd885-5fc1-4bc8-9b47-57e3103e9197" containerName="mariadb-database-create" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.593850 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.596090 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m498w" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.600248 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.600624 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.600325 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.604691 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.606545 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.609186 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.612160 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.621704 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.633340 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.714834 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.715507 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.715633 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.715755 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-scripts\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.715866 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.715969 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-logs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716120 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716252 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgff5\" (UniqueName: \"kubernetes.io/projected/23c6ec92-8230-4322-ad9c-67c0a3260eec-kube-api-access-mgff5\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716368 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716471 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716559 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716665 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-config-data\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716882 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bj7h\" (UniqueName: \"kubernetes.io/projected/4454a179-0fc2-4623-b1ba-b7946009d769-kube-api-access-9bj7h\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.716991 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-logs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.717114 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.819066 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820140 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820311 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820409 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-config-data\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820528 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bj7h\" (UniqueName: \"kubernetes.io/projected/4454a179-0fc2-4623-b1ba-b7946009d769-kube-api-access-9bj7h\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820620 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-logs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820700 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820804 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.820927 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821014 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821098 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-scripts\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821181 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821976 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-logs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.822060 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.822197 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgff5\" (UniqueName: \"kubernetes.io/projected/23c6ec92-8230-4322-ad9c-67c0a3260eec-kube-api-access-mgff5\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.823041 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.824307 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.821282 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.828642 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.828942 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c6ec92-8230-4322-ad9c-67c0a3260eec-logs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.829141 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.829968 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.830596 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4454a179-0fc2-4623-b1ba-b7946009d769-logs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.841009 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-config-data\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.848785 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.849243 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-scripts\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.849769 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4454a179-0fc2-4623-b1ba-b7946009d769-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.853441 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.854069 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c6ec92-8230-4322-ad9c-67c0a3260eec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.863464 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgff5\" (UniqueName: \"kubernetes.io/projected/23c6ec92-8230-4322-ad9c-67c0a3260eec-kube-api-access-mgff5\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.885473 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bj7h\" (UniqueName: \"kubernetes.io/projected/4454a179-0fc2-4623-b1ba-b7946009d769-kube-api-access-9bj7h\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.904443 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"4454a179-0fc2-4623-b1ba-b7946009d769\") " pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.914286 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"23c6ec92-8230-4322-ad9c-67c0a3260eec\") " pod="openstack/glance-default-external-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.916491 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:19 crc kubenswrapper[4847]: I1210 14:44:19.929090 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.676583 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.779772 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b67bb20-a54d-4e08-8d05-c68fd77bd38f" path="/var/lib/kubelet/pods/6b67bb20-a54d-4e08-8d05-c68fd77bd38f/volumes" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.780486 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="accad8cb-ba38-44f0-bbac-20c879e8a2b9" path="/var/lib/kubelet/pods/accad8cb-ba38-44f0-bbac-20c879e8a2b9/volumes" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.860358 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.933545 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lgckt"] Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.934962 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.941359 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.941638 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ww4k6" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.941911 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 14:44:20 crc kubenswrapper[4847]: I1210 14:44:20.947323 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lgckt"] Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.048680 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.049079 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.049141 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.049187 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlgcv\" (UniqueName: \"kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.150748 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.150824 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlgcv\" (UniqueName: \"kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.150874 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.150928 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.175681 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.175875 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.183804 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlgcv\" (UniqueName: \"kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.185869 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data\") pod \"nova-cell0-conductor-db-sync-lgckt\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.284734 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.505935 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23c6ec92-8230-4322-ad9c-67c0a3260eec","Type":"ContainerStarted","Data":"a775e244a3844282dc2db68bd1aa67b314d9debe388a5e10e7cee17bb4d467e1"} Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.513592 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerStarted","Data":"453192b47f5ec4afa58bf770c0d8f6e6bc439b56cabd0f647a638419e8a629a7"} Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.526319 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4454a179-0fc2-4623-b1ba-b7946009d769","Type":"ContainerStarted","Data":"643622445d5ae5d50bb1229afddcb185bdbe554a99eff61d49091946b7f143d0"} Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.850587 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-286040590bf3f27ab9fe7d49b541dc9b72920dad78bf336a96c4f782df5e40fd": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-286040590bf3f27ab9fe7d49b541dc9b72920dad78bf336a96c4f782df5e40fd: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: I1210 14:44:21.851528 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lgckt"] Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.866081 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ad03ce_c5f0_4ba1_8f44_bf7049dee1f6.slice/crio-b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ad03ce_c5f0_4ba1_8f44_bf7049dee1f6.slice/crio-b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.866129 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a1d69_c6b6_434e_b0e3_7c43c39a2e1f.slice/crio-conmon-2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a1d69_c6b6_434e_b0e3_7c43c39a2e1f.slice/crio-conmon-2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910298 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4fd885_5fc1_4bc8_9b47_57e3103e9197.slice/crio-conmon-cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4fd885_5fc1_4bc8_9b47_57e3103e9197.slice/crio-conmon-cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910345 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a1d69_c6b6_434e_b0e3_7c43c39a2e1f.slice/crio-2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a1d69_c6b6_434e_b0e3_7c43c39a2e1f.slice/crio-2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910359 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4fd885_5fc1_4bc8_9b47_57e3103e9197.slice/crio-cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4fd885_5fc1_4bc8_9b47_57e3103e9197.slice/crio-cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910375 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod193125a6_9a88_47b0_b462_daab3aa61691.slice/crio-conmon-b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod193125a6_9a88_47b0_b462_daab3aa61691.slice/crio-conmon-b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910388 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-conmon-f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-conmon-f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910406 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod193125a6_9a88_47b0_b462_daab3aa61691.slice/crio-b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod193125a6_9a88_47b0_b462_daab3aa61691.slice/crio-b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68.scope: no such file or directory Dec 10 14:44:21 crc kubenswrapper[4847]: W1210 14:44:21.910418 4847 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice/crio-f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b.scope: no such file or directory Dec 10 14:44:22 crc kubenswrapper[4847]: E1210 14:44:22.178673 4847 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c9841e_2a1b_4796_99ca_cd6eeac1b5bc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a4fd885_5fc1_4bc8_9b47_57e3103e9197.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ad03ce_c5f0_4ba1_8f44_bf7049dee1f6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaccad8cb_ba38_44f0_bbac_20c879e8a2b9.slice/crio-conmon-109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b67bb20_a54d_4e08_8d05_c68fd77bd38f.slice/crio-5e55074c6e6d6f862dbb3be9c6bc04a1b178630bb58a096302288e4503ca500c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b67bb20_a54d_4e08_8d05_c68fd77bd38f.slice/crio-e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c9841e_2a1b_4796_99ca_cd6eeac1b5bc.slice/crio-f17376aed3e3b044cbe8a08d1867f4dd1be47d41042e41a14abe38516a4bd620\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaccad8cb_ba38_44f0_bbac_20c879e8a2b9.slice/crio-109e2370e5b9356f72ff0b343c7f0b0ed74c5a7e61c15fd409e4581c8d620729.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333a1d69_c6b6_434e_b0e3_7c43c39a2e1f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6c2e2c8_cc11_4db9_a572_4fb3e4e9c992.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod517d4777_6273_401e_bf27_b23ea11f1a41.slice/crio-d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b67bb20_a54d_4e08_8d05_c68fd77bd38f.slice/crio-b7dd675822e65013af0c5cf52a4152c8973693d09ef46baf9312fb75ab4a1640.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b67bb20_a54d_4e08_8d05_c68fd77bd38f.slice/crio-conmon-e2c227e881f26c1c3a2ff5ac39050724e0e2ce311504fb93df2f18609a0688c1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b6a1839_419a_4f80_9c45_ff37701bb916.slice\": RecentStats: unable to find data in memory cache]" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.283221 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.410729 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.410834 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7n5r\" (UniqueName: \"kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.410909 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.410958 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.410980 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.411034 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.411117 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs\") pod \"517d4777-6273-401e-bf27-b23ea11f1a41\" (UID: \"517d4777-6273-401e-bf27-b23ea11f1a41\") " Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.412376 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs" (OuterVolumeSpecName: "logs") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.415482 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.422117 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r" (OuterVolumeSpecName: "kube-api-access-r7n5r") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "kube-api-access-r7n5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.445888 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts" (OuterVolumeSpecName: "scripts") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.449960 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data" (OuterVolumeSpecName: "config-data") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.500165 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517153 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517210 4847 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517220 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517229 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/517d4777-6273-401e-bf27-b23ea11f1a41-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517238 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517d4777-6273-401e-bf27-b23ea11f1a41-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.517245 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7n5r\" (UniqueName: \"kubernetes.io/projected/517d4777-6273-401e-bf27-b23ea11f1a41-kube-api-access-r7n5r\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.527667 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "517d4777-6273-401e-bf27-b23ea11f1a41" (UID: "517d4777-6273-401e-bf27-b23ea11f1a41"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.542837 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4454a179-0fc2-4623-b1ba-b7946009d769","Type":"ContainerStarted","Data":"6f04324faf17c93bff06f5ae0c0de10de3144e1e655c681ac2f6349f4485e622"} Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.554112 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lgckt" event={"ID":"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908","Type":"ContainerStarted","Data":"3e3fec85e4de5ae67e4345e6c9929f9198351a3e3e2ecbb621e5bca1d26834b1"} Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.562431 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23c6ec92-8230-4322-ad9c-67c0a3260eec","Type":"ContainerStarted","Data":"d48b7ca60ef966ca1ede4053696fe5f3fe09fd38643c8293f100486821892fd2"} Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.565327 4847 generic.go:334] "Generic (PLEG): container finished" podID="517d4777-6273-401e-bf27-b23ea11f1a41" containerID="d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f" exitCode=137 Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.565362 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerDied","Data":"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f"} Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.565383 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8676865fc8-rvrhh" event={"ID":"517d4777-6273-401e-bf27-b23ea11f1a41","Type":"ContainerDied","Data":"7a1ce065e6d8ac47fefffb536468e2b50c7ad3d69aae55948b0e351f522b65b3"} Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.565404 4847 scope.go:117] "RemoveContainer" containerID="54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.565541 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8676865fc8-rvrhh" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.615753 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.619023 4847 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/517d4777-6273-401e-bf27-b23ea11f1a41-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.620828 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8676865fc8-rvrhh"] Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.776738 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" path="/var/lib/kubelet/pods/517d4777-6273-401e-bf27-b23ea11f1a41/volumes" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.815600 4847 scope.go:117] "RemoveContainer" containerID="d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.852993 4847 scope.go:117] "RemoveContainer" containerID="54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50" Dec 10 14:44:22 crc kubenswrapper[4847]: E1210 14:44:22.872553 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50\": container with ID starting with 54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50 not found: ID does not exist" containerID="54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.872612 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50"} err="failed to get container status \"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50\": rpc error: code = NotFound desc = could not find container \"54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50\": container with ID starting with 54746c1f5c4154dd0cd732a519a1e56e9e118680a2b9532f553ed85136e52f50 not found: ID does not exist" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.872645 4847 scope.go:117] "RemoveContainer" containerID="d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f" Dec 10 14:44:22 crc kubenswrapper[4847]: E1210 14:44:22.873251 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f\": container with ID starting with d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f not found: ID does not exist" containerID="d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.873290 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f"} err="failed to get container status \"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f\": rpc error: code = NotFound desc = could not find container \"d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f\": container with ID starting with d80eb16e5dafbf5e4f3393748136519f27b91879fd94136d1dd3236a2db3ca0f not found: ID does not exist" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.878968 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:22 crc kubenswrapper[4847]: I1210 14:44:22.899928 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65d99c95b7-t6kg4" Dec 10 14:44:23 crc kubenswrapper[4847]: I1210 14:44:23.594329 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4454a179-0fc2-4623-b1ba-b7946009d769","Type":"ContainerStarted","Data":"4068eda7e62d2957ab7681ff54398417ad2b1473a89958f0b8dc394a07af8ccf"} Dec 10 14:44:23 crc kubenswrapper[4847]: I1210 14:44:23.599797 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"23c6ec92-8230-4322-ad9c-67c0a3260eec","Type":"ContainerStarted","Data":"633bf9f62f9d8c785ebe5bffa04d7abc54b1f43b530f65f752e7db94b519f3b8"} Dec 10 14:44:23 crc kubenswrapper[4847]: I1210 14:44:23.720937 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.720917104 podStartE2EDuration="4.720917104s" podCreationTimestamp="2025-12-10 14:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:23.634320484 +0000 UTC m=+1213.203538134" watchObservedRunningTime="2025-12-10 14:44:23.720917104 +0000 UTC m=+1213.290134724" Dec 10 14:44:23 crc kubenswrapper[4847]: I1210 14:44:23.722025 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.7220173039999995 podStartE2EDuration="4.722017304s" podCreationTimestamp="2025-12-10 14:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:23.696362186 +0000 UTC m=+1213.265579836" watchObservedRunningTime="2025-12-10 14:44:23.722017304 +0000 UTC m=+1213.291234934" Dec 10 14:44:24 crc kubenswrapper[4847]: I1210 14:44:24.623016 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerStarted","Data":"b0e0b47b985bcbac4607e609ba19a83eb84c08b88f043794f09f7c0ca1b049f9"} Dec 10 14:44:24 crc kubenswrapper[4847]: I1210 14:44:24.623287 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-central-agent" containerID="cri-o://99e9bbf8f3cf2525103e5a294714dcef51853c25b76f10c5e874dfcb410b6741" gracePeriod=30 Dec 10 14:44:24 crc kubenswrapper[4847]: I1210 14:44:24.623332 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="proxy-httpd" containerID="cri-o://b0e0b47b985bcbac4607e609ba19a83eb84c08b88f043794f09f7c0ca1b049f9" gracePeriod=30 Dec 10 14:44:24 crc kubenswrapper[4847]: I1210 14:44:24.623403 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-notification-agent" containerID="cri-o://fa6064489107e11585425f3a26327327c3ec986550ca11e1d165d6f6ba9eb6bd" gracePeriod=30 Dec 10 14:44:24 crc kubenswrapper[4847]: I1210 14:44:24.623372 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="sg-core" containerID="cri-o://453192b47f5ec4afa58bf770c0d8f6e6bc439b56cabd0f647a638419e8a629a7" gracePeriod=30 Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.646731 4847 generic.go:334] "Generic (PLEG): container finished" podID="696b0780-bd3e-480e-a338-711b34dc6851" containerID="b0e0b47b985bcbac4607e609ba19a83eb84c08b88f043794f09f7c0ca1b049f9" exitCode=0 Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647069 4847 generic.go:334] "Generic (PLEG): container finished" podID="696b0780-bd3e-480e-a338-711b34dc6851" containerID="453192b47f5ec4afa58bf770c0d8f6e6bc439b56cabd0f647a638419e8a629a7" exitCode=2 Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647096 4847 generic.go:334] "Generic (PLEG): container finished" podID="696b0780-bd3e-480e-a338-711b34dc6851" containerID="fa6064489107e11585425f3a26327327c3ec986550ca11e1d165d6f6ba9eb6bd" exitCode=0 Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647102 4847 generic.go:334] "Generic (PLEG): container finished" podID="696b0780-bd3e-480e-a338-711b34dc6851" containerID="99e9bbf8f3cf2525103e5a294714dcef51853c25b76f10c5e874dfcb410b6741" exitCode=0 Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647134 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerDied","Data":"b0e0b47b985bcbac4607e609ba19a83eb84c08b88f043794f09f7c0ca1b049f9"} Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647212 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerDied","Data":"453192b47f5ec4afa58bf770c0d8f6e6bc439b56cabd0f647a638419e8a629a7"} Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647223 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerDied","Data":"fa6064489107e11585425f3a26327327c3ec986550ca11e1d165d6f6ba9eb6bd"} Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.647231 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerDied","Data":"99e9bbf8f3cf2525103e5a294714dcef51853c25b76f10c5e874dfcb410b6741"} Dec 10 14:44:25 crc kubenswrapper[4847]: I1210 14:44:25.888776 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.011690 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012070 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzflw\" (UniqueName: \"kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012221 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012302 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012350 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012425 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.012499 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd\") pod \"696b0780-bd3e-480e-a338-711b34dc6851\" (UID: \"696b0780-bd3e-480e-a338-711b34dc6851\") " Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.013619 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.013966 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.019982 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw" (OuterVolumeSpecName: "kube-api-access-tzflw") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "kube-api-access-tzflw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.036888 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts" (OuterVolumeSpecName: "scripts") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.048947 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.095543 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115120 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115368 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115489 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115599 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/696b0780-bd3e-480e-a338-711b34dc6851-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115746 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.115849 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzflw\" (UniqueName: \"kubernetes.io/projected/696b0780-bd3e-480e-a338-711b34dc6851-kube-api-access-tzflw\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.140994 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data" (OuterVolumeSpecName: "config-data") pod "696b0780-bd3e-480e-a338-711b34dc6851" (UID: "696b0780-bd3e-480e-a338-711b34dc6851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.218237 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/696b0780-bd3e-480e-a338-711b34dc6851-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.659569 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"696b0780-bd3e-480e-a338-711b34dc6851","Type":"ContainerDied","Data":"b982a13b755b01c19f6f1a4da8dfee6513461a85a746d96bd585a49e7969f753"} Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.659629 4847 scope.go:117] "RemoveContainer" containerID="b0e0b47b985bcbac4607e609ba19a83eb84c08b88f043794f09f7c0ca1b049f9" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.659633 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.706515 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.725811 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739127 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739607 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="sg-core" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739627 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="sg-core" Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739650 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739657 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739668 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-central-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739675 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-central-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739690 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-notification-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739696 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-notification-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739722 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="proxy-httpd" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739730 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="proxy-httpd" Dec 10 14:44:26 crc kubenswrapper[4847]: E1210 14:44:26.739742 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon-log" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739748 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon-log" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739936 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-central-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739953 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon-log" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739963 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="sg-core" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739977 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="proxy-httpd" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739988 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="696b0780-bd3e-480e-a338-711b34dc6851" containerName="ceilometer-notification-agent" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.739996 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="517d4777-6273-401e-bf27-b23ea11f1a41" containerName="horizon" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.741621 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.744560 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.744832 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.758264 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.798800 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="696b0780-bd3e-480e-a338-711b34dc6851" path="/var/lib/kubelet/pods/696b0780-bd3e-480e-a338-711b34dc6851/volumes" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829596 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829744 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829777 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829806 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829874 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.829900 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.830138 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9pf9\" (UniqueName: \"kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932183 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9pf9\" (UniqueName: \"kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932303 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932390 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932417 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932470 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932823 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.932868 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.934371 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.934977 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.943305 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.944594 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.946243 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.955665 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9pf9\" (UniqueName: \"kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:26 crc kubenswrapper[4847]: I1210 14:44:26.956372 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts\") pod \"ceilometer-0\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " pod="openstack/ceilometer-0" Dec 10 14:44:27 crc kubenswrapper[4847]: I1210 14:44:27.139861 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.917390 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.917639 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.929448 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.929879 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.953085 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.970641 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.986565 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 14:44:29 crc kubenswrapper[4847]: I1210 14:44:29.986700 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 14:44:30 crc kubenswrapper[4847]: I1210 14:44:30.452474 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:30 crc kubenswrapper[4847]: I1210 14:44:30.709533 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 14:44:30 crc kubenswrapper[4847]: I1210 14:44:30.709593 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:30 crc kubenswrapper[4847]: I1210 14:44:30.709609 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:30 crc kubenswrapper[4847]: I1210 14:44:30.709625 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 14:44:32 crc kubenswrapper[4847]: I1210 14:44:32.478378 4847 scope.go:117] "RemoveContainer" containerID="453192b47f5ec4afa58bf770c0d8f6e6bc439b56cabd0f647a638419e8a629a7" Dec 10 14:44:32 crc kubenswrapper[4847]: I1210 14:44:32.583013 4847 scope.go:117] "RemoveContainer" containerID="fa6064489107e11585425f3a26327327c3ec986550ca11e1d165d6f6ba9eb6bd" Dec 10 14:44:32 crc kubenswrapper[4847]: I1210 14:44:32.762273 4847 scope.go:117] "RemoveContainer" containerID="99e9bbf8f3cf2525103e5a294714dcef51853c25b76f10c5e874dfcb410b6741" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.111355 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:33 crc kubenswrapper[4847]: W1210 14:44:33.116677 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33b1bd32_4a3a_47a6_abbf_f999ec1deed1.slice/crio-4bdf8c7dfe3a76a203b444f78e87e85d7d3a209e793f89f1cae56bf0e3352008 WatchSource:0}: Error finding container 4bdf8c7dfe3a76a203b444f78e87e85d7d3a209e793f89f1cae56bf0e3352008: Status 404 returned error can't find the container with id 4bdf8c7dfe3a76a203b444f78e87e85d7d3a209e793f89f1cae56bf0e3352008 Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.344436 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.344549 4847 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.350267 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.661431 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.661957 4847 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.763309 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lgckt" event={"ID":"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908","Type":"ContainerStarted","Data":"4f0d3ca62098db3f83d740b5799f740e768d4f35515a52f4dbc37dbe4955d797"} Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.765649 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerStarted","Data":"4bdf8c7dfe3a76a203b444f78e87e85d7d3a209e793f89f1cae56bf0e3352008"} Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.802441 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-lgckt" podStartSLOduration=3.110117067 podStartE2EDuration="13.802389212s" podCreationTimestamp="2025-12-10 14:44:20 +0000 UTC" firstStartedPulling="2025-12-10 14:44:21.894661531 +0000 UTC m=+1211.463879161" lastFinishedPulling="2025-12-10 14:44:32.586933676 +0000 UTC m=+1222.156151306" observedRunningTime="2025-12-10 14:44:33.781219928 +0000 UTC m=+1223.350437568" watchObservedRunningTime="2025-12-10 14:44:33.802389212 +0000 UTC m=+1223.371606852" Dec 10 14:44:33 crc kubenswrapper[4847]: I1210 14:44:33.888380 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 14:44:34 crc kubenswrapper[4847]: I1210 14:44:34.790384 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerStarted","Data":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} Dec 10 14:44:35 crc kubenswrapper[4847]: I1210 14:44:35.818924 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerStarted","Data":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} Dec 10 14:44:35 crc kubenswrapper[4847]: I1210 14:44:35.819338 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerStarted","Data":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.839640 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerStarted","Data":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.840243 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.839840 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-central-agent" containerID="cri-o://cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" gracePeriod=30 Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.840062 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-notification-agent" containerID="cri-o://a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" gracePeriod=30 Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.840090 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="sg-core" containerID="cri-o://c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" gracePeriod=30 Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.839921 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="proxy-httpd" containerID="cri-o://4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" gracePeriod=30 Dec 10 14:44:37 crc kubenswrapper[4847]: I1210 14:44:37.882357 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.167516339 podStartE2EDuration="11.882328695s" podCreationTimestamp="2025-12-10 14:44:26 +0000 UTC" firstStartedPulling="2025-12-10 14:44:33.12369018 +0000 UTC m=+1222.692907810" lastFinishedPulling="2025-12-10 14:44:36.838502536 +0000 UTC m=+1226.407720166" observedRunningTime="2025-12-10 14:44:37.870628842 +0000 UTC m=+1227.439846472" watchObservedRunningTime="2025-12-10 14:44:37.882328695 +0000 UTC m=+1227.451546325" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.781005 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851764 4847 generic.go:334] "Generic (PLEG): container finished" podID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" exitCode=0 Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851809 4847 generic.go:334] "Generic (PLEG): container finished" podID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" exitCode=2 Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851821 4847 generic.go:334] "Generic (PLEG): container finished" podID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" exitCode=0 Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851830 4847 generic.go:334] "Generic (PLEG): container finished" podID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" exitCode=0 Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851859 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerDied","Data":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerDied","Data":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851913 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerDied","Data":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851923 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerDied","Data":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851934 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"33b1bd32-4a3a-47a6-abbf-f999ec1deed1","Type":"ContainerDied","Data":"4bdf8c7dfe3a76a203b444f78e87e85d7d3a209e793f89f1cae56bf0e3352008"} Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.851957 4847 scope.go:117] "RemoveContainer" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.852189 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.874820 4847 scope.go:117] "RemoveContainer" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.889440 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.889601 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.889812 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.889875 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.890119 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.890167 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9pf9\" (UniqueName: \"kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.890241 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data\") pod \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\" (UID: \"33b1bd32-4a3a-47a6-abbf-f999ec1deed1\") " Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.891660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.893063 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.894821 4847 scope.go:117] "RemoveContainer" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.896035 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts" (OuterVolumeSpecName: "scripts") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.897854 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9" (OuterVolumeSpecName: "kube-api-access-k9pf9") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "kube-api-access-k9pf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.916473 4847 scope.go:117] "RemoveContainer" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.920395 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.969571 4847 scope.go:117] "RemoveContainer" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: E1210 14:44:38.970097 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": container with ID starting with 4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b not found: ID does not exist" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.970153 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} err="failed to get container status \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": rpc error: code = NotFound desc = could not find container \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": container with ID starting with 4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.970190 4847 scope.go:117] "RemoveContainer" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: E1210 14:44:38.970538 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": container with ID starting with c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c not found: ID does not exist" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.970571 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} err="failed to get container status \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": rpc error: code = NotFound desc = could not find container \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": container with ID starting with c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.970568 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.970591 4847 scope.go:117] "RemoveContainer" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: E1210 14:44:38.970970 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": container with ID starting with a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b not found: ID does not exist" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971019 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} err="failed to get container status \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": rpc error: code = NotFound desc = could not find container \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": container with ID starting with a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971067 4847 scope.go:117] "RemoveContainer" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: E1210 14:44:38.971453 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": container with ID starting with cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06 not found: ID does not exist" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971481 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} err="failed to get container status \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": rpc error: code = NotFound desc = could not find container \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": container with ID starting with cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06 not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971497 4847 scope.go:117] "RemoveContainer" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971767 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} err="failed to get container status \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": rpc error: code = NotFound desc = could not find container \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": container with ID starting with 4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.971792 4847 scope.go:117] "RemoveContainer" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972039 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} err="failed to get container status \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": rpc error: code = NotFound desc = could not find container \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": container with ID starting with c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972067 4847 scope.go:117] "RemoveContainer" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972258 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} err="failed to get container status \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": rpc error: code = NotFound desc = could not find container \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": container with ID starting with a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972284 4847 scope.go:117] "RemoveContainer" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972514 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} err="failed to get container status \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": rpc error: code = NotFound desc = could not find container \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": container with ID starting with cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06 not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.972904 4847 scope.go:117] "RemoveContainer" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973152 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} err="failed to get container status \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": rpc error: code = NotFound desc = could not find container \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": container with ID starting with 4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973179 4847 scope.go:117] "RemoveContainer" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973541 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} err="failed to get container status \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": rpc error: code = NotFound desc = could not find container \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": container with ID starting with c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973560 4847 scope.go:117] "RemoveContainer" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973780 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} err="failed to get container status \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": rpc error: code = NotFound desc = could not find container \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": container with ID starting with a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.973803 4847 scope.go:117] "RemoveContainer" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974072 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} err="failed to get container status \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": rpc error: code = NotFound desc = could not find container \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": container with ID starting with cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06 not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974102 4847 scope.go:117] "RemoveContainer" containerID="4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974305 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b"} err="failed to get container status \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": rpc error: code = NotFound desc = could not find container \"4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b\": container with ID starting with 4c22230e0429bf762d862f45c167ff64cd434f018bec7bf532fac148136ab78b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974328 4847 scope.go:117] "RemoveContainer" containerID="c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974507 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c"} err="failed to get container status \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": rpc error: code = NotFound desc = could not find container \"c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c\": container with ID starting with c56a128b4c2ebd0e5aa99893ba6542cc16019f8817d0ff74c625c5b461a8bd8c not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974526 4847 scope.go:117] "RemoveContainer" containerID="a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974950 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b"} err="failed to get container status \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": rpc error: code = NotFound desc = could not find container \"a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b\": container with ID starting with a5aa1e583b7c4747067f9aaf9f023b1844aedbe92cdf94dcee047115eeda885b not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.974976 4847 scope.go:117] "RemoveContainer" containerID="cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.975202 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06"} err="failed to get container status \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": rpc error: code = NotFound desc = could not find container \"cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06\": container with ID starting with cfd9bd017638f05f0eda64476591d8144c5bcba83fe21df7c19883a0cf8e4f06 not found: ID does not exist" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993509 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993547 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993560 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993573 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993587 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9pf9\" (UniqueName: \"kubernetes.io/projected/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-kube-api-access-k9pf9\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:38 crc kubenswrapper[4847]: I1210 14:44:38.993599 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.007502 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data" (OuterVolumeSpecName: "config-data") pod "33b1bd32-4a3a-47a6-abbf-f999ec1deed1" (UID: "33b1bd32-4a3a-47a6-abbf-f999ec1deed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.095858 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b1bd32-4a3a-47a6-abbf-f999ec1deed1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.192774 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.206881 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.219922 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:39 crc kubenswrapper[4847]: E1210 14:44:39.220643 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-notification-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.220675 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-notification-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: E1210 14:44:39.220693 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="sg-core" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.220703 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="sg-core" Dec 10 14:44:39 crc kubenswrapper[4847]: E1210 14:44:39.220784 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-central-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.220795 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-central-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: E1210 14:44:39.220837 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="proxy-httpd" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.220849 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="proxy-httpd" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.221119 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-notification-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.221147 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="proxy-httpd" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.221171 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="sg-core" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.221183 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" containerName="ceilometer-central-agent" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.234877 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.235599 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.238038 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.238499 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402136 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402192 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402257 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402316 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7dqp\" (UniqueName: \"kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402395 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402418 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.402434 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503699 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503766 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503785 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503840 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503865 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503910 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.503945 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7dqp\" (UniqueName: \"kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.504286 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.505177 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.508946 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.508981 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.511624 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.512205 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.524359 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7dqp\" (UniqueName: \"kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp\") pod \"ceilometer-0\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " pod="openstack/ceilometer-0" Dec 10 14:44:39 crc kubenswrapper[4847]: I1210 14:44:39.559556 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:44:40 crc kubenswrapper[4847]: I1210 14:44:40.056814 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:44:40 crc kubenswrapper[4847]: W1210 14:44:40.063176 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc74ba68c_4c24_488f_a863_e35cece9f1d9.slice/crio-eb6f05822d47bbe0a768b5e175dd7ab7a0405792fd73958aef1a0ae2487b58b9 WatchSource:0}: Error finding container eb6f05822d47bbe0a768b5e175dd7ab7a0405792fd73958aef1a0ae2487b58b9: Status 404 returned error can't find the container with id eb6f05822d47bbe0a768b5e175dd7ab7a0405792fd73958aef1a0ae2487b58b9 Dec 10 14:44:40 crc kubenswrapper[4847]: I1210 14:44:40.772502 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33b1bd32-4a3a-47a6-abbf-f999ec1deed1" path="/var/lib/kubelet/pods/33b1bd32-4a3a-47a6-abbf-f999ec1deed1/volumes" Dec 10 14:44:40 crc kubenswrapper[4847]: I1210 14:44:40.873408 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerStarted","Data":"eb6f05822d47bbe0a768b5e175dd7ab7a0405792fd73958aef1a0ae2487b58b9"} Dec 10 14:44:42 crc kubenswrapper[4847]: I1210 14:44:42.892754 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerStarted","Data":"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec"} Dec 10 14:44:42 crc kubenswrapper[4847]: I1210 14:44:42.893825 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerStarted","Data":"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1"} Dec 10 14:44:44 crc kubenswrapper[4847]: I1210 14:44:44.916736 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerStarted","Data":"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9"} Dec 10 14:44:46 crc kubenswrapper[4847]: I1210 14:44:46.933670 4847 generic.go:334] "Generic (PLEG): container finished" podID="56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" containerID="4f0d3ca62098db3f83d740b5799f740e768d4f35515a52f4dbc37dbe4955d797" exitCode=0 Dec 10 14:44:46 crc kubenswrapper[4847]: I1210 14:44:46.933783 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lgckt" event={"ID":"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908","Type":"ContainerDied","Data":"4f0d3ca62098db3f83d740b5799f740e768d4f35515a52f4dbc37dbe4955d797"} Dec 10 14:44:46 crc kubenswrapper[4847]: I1210 14:44:46.943464 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerStarted","Data":"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5"} Dec 10 14:44:46 crc kubenswrapper[4847]: I1210 14:44:46.943812 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:44:46 crc kubenswrapper[4847]: I1210 14:44:46.978810 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.3238211 podStartE2EDuration="7.978791594s" podCreationTimestamp="2025-12-10 14:44:39 +0000 UTC" firstStartedPulling="2025-12-10 14:44:40.066210998 +0000 UTC m=+1229.635428628" lastFinishedPulling="2025-12-10 14:44:45.721181492 +0000 UTC m=+1235.290399122" observedRunningTime="2025-12-10 14:44:46.972256494 +0000 UTC m=+1236.541474124" watchObservedRunningTime="2025-12-10 14:44:46.978791594 +0000 UTC m=+1236.548009224" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.274857 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.363241 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle\") pod \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.363288 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts\") pod \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.363409 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlgcv\" (UniqueName: \"kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv\") pod \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.363544 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data\") pod \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\" (UID: \"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908\") " Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.369566 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts" (OuterVolumeSpecName: "scripts") pod "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" (UID: "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.370105 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv" (OuterVolumeSpecName: "kube-api-access-vlgcv") pod "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" (UID: "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908"). InnerVolumeSpecName "kube-api-access-vlgcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.393188 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data" (OuterVolumeSpecName: "config-data") pod "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" (UID: "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.406260 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" (UID: "56f3fb84-6bbc-478c-9f5c-a6adcd0b1908"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.465975 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.466210 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.466268 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.466321 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlgcv\" (UniqueName: \"kubernetes.io/projected/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908-kube-api-access-vlgcv\") on node \"crc\" DevicePath \"\"" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.960244 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-lgckt" event={"ID":"56f3fb84-6bbc-478c-9f5c-a6adcd0b1908","Type":"ContainerDied","Data":"3e3fec85e4de5ae67e4345e6c9929f9198351a3e3e2ecbb621e5bca1d26834b1"} Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.960525 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e3fec85e4de5ae67e4345e6c9929f9198351a3e3e2ecbb621e5bca1d26834b1" Dec 10 14:44:48 crc kubenswrapper[4847]: I1210 14:44:48.960651 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-lgckt" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.057931 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 14:44:49 crc kubenswrapper[4847]: E1210 14:44:49.058560 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" containerName="nova-cell0-conductor-db-sync" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.058576 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" containerName="nova-cell0-conductor-db-sync" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.058770 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" containerName="nova-cell0-conductor-db-sync" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.059342 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.061400 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ww4k6" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.063189 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.072654 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.181478 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t44t8\" (UniqueName: \"kubernetes.io/projected/e9315ad6-b2ce-420a-b8a4-c168264662de-kube-api-access-t44t8\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.181574 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.181663 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.283877 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.283947 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.284069 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t44t8\" (UniqueName: \"kubernetes.io/projected/e9315ad6-b2ce-420a-b8a4-c168264662de-kube-api-access-t44t8\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.289410 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.289598 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9315ad6-b2ce-420a-b8a4-c168264662de-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.303029 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t44t8\" (UniqueName: \"kubernetes.io/projected/e9315ad6-b2ce-420a-b8a4-c168264662de-kube-api-access-t44t8\") pod \"nova-cell0-conductor-0\" (UID: \"e9315ad6-b2ce-420a-b8a4-c168264662de\") " pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.387639 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.819468 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 14:44:49 crc kubenswrapper[4847]: W1210 14:44:49.820169 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9315ad6_b2ce_420a_b8a4_c168264662de.slice/crio-7479102ac263752157835ae6ad9cc75dde407829820818d2f9c960ce3c5ff9c8 WatchSource:0}: Error finding container 7479102ac263752157835ae6ad9cc75dde407829820818d2f9c960ce3c5ff9c8: Status 404 returned error can't find the container with id 7479102ac263752157835ae6ad9cc75dde407829820818d2f9c960ce3c5ff9c8 Dec 10 14:44:49 crc kubenswrapper[4847]: I1210 14:44:49.969921 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e9315ad6-b2ce-420a-b8a4-c168264662de","Type":"ContainerStarted","Data":"7479102ac263752157835ae6ad9cc75dde407829820818d2f9c960ce3c5ff9c8"} Dec 10 14:44:50 crc kubenswrapper[4847]: I1210 14:44:50.980699 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e9315ad6-b2ce-420a-b8a4-c168264662de","Type":"ContainerStarted","Data":"8af5ba5220cb0fe09e381d2189e8ed044caf802806740e1b83658e17b36446fd"} Dec 10 14:44:50 crc kubenswrapper[4847]: I1210 14:44:50.980938 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:51 crc kubenswrapper[4847]: I1210 14:44:51.001406 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.001379662 podStartE2EDuration="2.001379662s" podCreationTimestamp="2025-12-10 14:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:50.993511907 +0000 UTC m=+1240.562729527" watchObservedRunningTime="2025-12-10 14:44:51.001379662 +0000 UTC m=+1240.570597292" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.414602 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.832694 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zdcv2"] Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.834188 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.836144 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.836297 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.847419 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zdcv2"] Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.883333 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hsbm\" (UniqueName: \"kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.883681 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.884010 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.884153 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.976569 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.978296 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.981004 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.986428 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.986529 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.986571 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.986668 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hsbm\" (UniqueName: \"kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:54 crc kubenswrapper[4847]: I1210 14:44:54.991770 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.007583 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.007798 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.019394 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.023369 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hsbm\" (UniqueName: \"kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm\") pod \"nova-cell0-cell-mapping-zdcv2\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.095407 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.095447 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5b87\" (UniqueName: \"kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.095466 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.099990 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.101620 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.123043 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.145490 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.159913 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.160370 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.161641 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.164725 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.185497 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.196778 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.196921 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp7ct\" (UniqueName: \"kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.197069 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.197135 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.197178 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.197207 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5b87\" (UniqueName: \"kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.197228 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.207424 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.210601 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.230480 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5b87\" (UniqueName: \"kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87\") pod \"nova-cell1-novncproxy-0\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.284134 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.286337 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.294330 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316194 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316465 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316537 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp7ct\" (UniqueName: \"kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316595 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrvq\" (UniqueName: \"kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316630 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316817 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.316974 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.317332 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.333029 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.338422 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.339542 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp7ct\" (UniqueName: \"kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.358626 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data\") pod \"nova-api-0\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.385370 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.387360 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.421991 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29rdp\" (UniqueName: \"kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422086 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422125 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422278 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422318 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrvq\" (UniqueName: \"kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422373 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.422458 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.426488 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.456446 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrvq\" (UniqueName: \"kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.457186 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.469658 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.471371 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data\") pod \"nova-scheduler-0\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.478562 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.484000 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.524823 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525002 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525059 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525359 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525495 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29rdp\" (UniqueName: \"kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525591 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525786 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.525864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8bk\" (UniqueName: \"kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.526939 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.527015 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.527679 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.533411 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.539187 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.551157 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29rdp\" (UniqueName: \"kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp\") pod \"nova-metadata-0\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629316 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8bk\" (UniqueName: \"kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629372 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629409 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629454 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629480 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.629532 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.630560 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.630686 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.631303 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.631292 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.634883 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.651336 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8bk\" (UniqueName: \"kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk\") pod \"dnsmasq-dns-bccf8f775-wwn4z\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.688337 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.722183 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.857356 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zdcv2"] Dec 10 14:44:55 crc kubenswrapper[4847]: I1210 14:44:55.984490 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.060970 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zdcv2" event={"ID":"5dd1e94d-67a9-443c-9870-81a88bf56445","Type":"ContainerStarted","Data":"ac82dc16abaae323f94a244ad6b13ae4cd8d7bc5c4acc146a13015c23cd17629"} Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.064767 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03","Type":"ContainerStarted","Data":"b45bbfd19b8adcf4cee809ed5f3ccc230e7647a213ae49c394e88d94c5270845"} Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.111613 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.159493 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8bgjb"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.178496 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8bgjb"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.178641 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.181780 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.182358 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.240123 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.240611 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zflb\" (UniqueName: \"kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.241002 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.241268 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.243250 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.332340 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:44:56 crc kubenswrapper[4847]: W1210 14:44:56.340913 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff251a94_9385_4610_a270_c9dea5c56272.slice/crio-0eb9584970868baf062b89fb11040a0e7a6e26dba68009b246d351963e29d4a9 WatchSource:0}: Error finding container 0eb9584970868baf062b89fb11040a0e7a6e26dba68009b246d351963e29d4a9: Status 404 returned error can't find the container with id 0eb9584970868baf062b89fb11040a0e7a6e26dba68009b246d351963e29d4a9 Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.343736 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.343828 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zflb\" (UniqueName: \"kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.343947 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.344052 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.348491 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.350289 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.355924 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.361611 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.365528 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zflb\" (UniqueName: \"kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb\") pod \"nova-cell1-conductor-db-sync-8bgjb\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:56 crc kubenswrapper[4847]: W1210 14:44:56.373196 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod037939fa_5737_4f3e_8a25_b7532272d5f1.slice/crio-af64a8b760f9323eb48d24d4c05b44d596840dc5ac4dff76cfd566fc3f59f99c WatchSource:0}: Error finding container af64a8b760f9323eb48d24d4c05b44d596840dc5ac4dff76cfd566fc3f59f99c: Status 404 returned error can't find the container with id af64a8b760f9323eb48d24d4c05b44d596840dc5ac4dff76cfd566fc3f59f99c Dec 10 14:44:56 crc kubenswrapper[4847]: I1210 14:44:56.519108 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.060976 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8bgjb"] Dec 10 14:44:57 crc kubenswrapper[4847]: W1210 14:44:57.066138 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0ae39e9_b424_481b_a316_403a59eb4837.slice/crio-15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb WatchSource:0}: Error finding container 15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb: Status 404 returned error can't find the container with id 15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.104172 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerStarted","Data":"0eb9584970868baf062b89fb11040a0e7a6e26dba68009b246d351963e29d4a9"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.113834 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zdcv2" event={"ID":"5dd1e94d-67a9-443c-9870-81a88bf56445","Type":"ContainerStarted","Data":"cd7aebf5817a579950b349a82d3860e10ad53f542e61f2fd453d39385993045d"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.135222 4847 generic.go:334] "Generic (PLEG): container finished" podID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerID="2f08f2148881fdb195190e1ef241a8177fd83a607a62c6a96844e52e4abaceca" exitCode=0 Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.135456 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" event={"ID":"037939fa-5737-4f3e-8a25-b7532272d5f1","Type":"ContainerDied","Data":"2f08f2148881fdb195190e1ef241a8177fd83a607a62c6a96844e52e4abaceca"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.135518 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" event={"ID":"037939fa-5737-4f3e-8a25-b7532272d5f1","Type":"ContainerStarted","Data":"af64a8b760f9323eb48d24d4c05b44d596840dc5ac4dff76cfd566fc3f59f99c"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.139911 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23c0a0d3-8946-4120-b757-d096b31f67c8","Type":"ContainerStarted","Data":"6ba6c4b1a827271949b0e4e4c3af499fe38f248d06e429c47bc8578f11ea7d97"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.144651 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerStarted","Data":"f10a7ce47b2c1973402880825983040ef7ab2431f121a8271ea941566cc95d0c"} Dec 10 14:44:57 crc kubenswrapper[4847]: I1210 14:44:57.213892 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zdcv2" podStartSLOduration=3.213868973 podStartE2EDuration="3.213868973s" podCreationTimestamp="2025-12-10 14:44:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:57.134773344 +0000 UTC m=+1246.703990984" watchObservedRunningTime="2025-12-10 14:44:57.213868973 +0000 UTC m=+1246.783086603" Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.159917 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" event={"ID":"037939fa-5737-4f3e-8a25-b7532272d5f1","Type":"ContainerStarted","Data":"d87007198c6c559570e1e0a3e0bb6b1acc336f37eb62d921a078a4c9a387f7df"} Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.160271 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.162500 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" event={"ID":"f0ae39e9-b424-481b-a316-403a59eb4837","Type":"ContainerStarted","Data":"68a84f74c916db4f5a2cbd1b3d639386df8b56b0b4509c4be4dde8a49eba44cf"} Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.162561 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" event={"ID":"f0ae39e9-b424-481b-a316-403a59eb4837","Type":"ContainerStarted","Data":"15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb"} Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.188082 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" podStartSLOduration=3.188015715 podStartE2EDuration="3.188015715s" podCreationTimestamp="2025-12-10 14:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:58.183656062 +0000 UTC m=+1247.752873702" watchObservedRunningTime="2025-12-10 14:44:58.188015715 +0000 UTC m=+1247.757233335" Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.222971 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" podStartSLOduration=2.222948825 podStartE2EDuration="2.222948825s" podCreationTimestamp="2025-12-10 14:44:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:44:58.196943718 +0000 UTC m=+1247.766161348" watchObservedRunningTime="2025-12-10 14:44:58.222948825 +0000 UTC m=+1247.792166455" Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.645124 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:44:58 crc kubenswrapper[4847]: I1210 14:44:58.656632 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.140781 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7"] Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.143100 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.145877 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.146292 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.162966 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7"] Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.226610 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerStarted","Data":"5a715d0deacfa5eab8ca5886fad6232acb9fc56851e5b57afeb867b087731eb4"} Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.228642 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0" gracePeriod=30 Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.229022 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03","Type":"ContainerStarted","Data":"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0"} Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.237666 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23c0a0d3-8946-4120-b757-d096b31f67c8","Type":"ContainerStarted","Data":"1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae"} Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.249831 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerStarted","Data":"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32"} Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.266096 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.266305 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lzc6\" (UniqueName: \"kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.266392 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.277646 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.644016718 podStartE2EDuration="6.277620258s" podCreationTimestamp="2025-12-10 14:44:54 +0000 UTC" firstStartedPulling="2025-12-10 14:44:55.985337579 +0000 UTC m=+1245.554555209" lastFinishedPulling="2025-12-10 14:44:59.618941119 +0000 UTC m=+1249.188158749" observedRunningTime="2025-12-10 14:45:00.264340532 +0000 UTC m=+1249.833558182" watchObservedRunningTime="2025-12-10 14:45:00.277620258 +0000 UTC m=+1249.846837888" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.292522 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.939433498 podStartE2EDuration="5.292495955s" podCreationTimestamp="2025-12-10 14:44:55 +0000 UTC" firstStartedPulling="2025-12-10 14:44:56.256815897 +0000 UTC m=+1245.826033527" lastFinishedPulling="2025-12-10 14:44:59.609878354 +0000 UTC m=+1249.179095984" observedRunningTime="2025-12-10 14:45:00.286949431 +0000 UTC m=+1249.856167061" watchObservedRunningTime="2025-12-10 14:45:00.292495955 +0000 UTC m=+1249.861713585" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.369100 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.369264 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lzc6\" (UniqueName: \"kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.369329 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.370837 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.376475 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.390688 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lzc6\" (UniqueName: \"kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6\") pod \"collect-profiles-29422965-pptl7\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.426987 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.485616 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 14:45:00 crc kubenswrapper[4847]: I1210 14:45:00.538411 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.100209 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7"] Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.276213 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" event={"ID":"183c9599-f0db-4e64-9ad1-3c4b60bef02c","Type":"ContainerStarted","Data":"2e45b9de4c72e243ea7edc7e2fb07620f6ba570f157c904f36bba4ce9f1fa325"} Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.282982 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerStarted","Data":"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449"} Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.300361 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-log" containerID="cri-o://5a715d0deacfa5eab8ca5886fad6232acb9fc56851e5b57afeb867b087731eb4" gracePeriod=30 Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.300627 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerStarted","Data":"af8d5747a7d81ff701f5b459dee8a83be0f14b7ce087be5bfb7ffe27ff7e60d5"} Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.300664 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-metadata" containerID="cri-o://af8d5747a7d81ff701f5b459dee8a83be0f14b7ce087be5bfb7ffe27ff7e60d5" gracePeriod=30 Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.315811 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.831733749 podStartE2EDuration="6.315785647s" podCreationTimestamp="2025-12-10 14:44:55 +0000 UTC" firstStartedPulling="2025-12-10 14:44:56.117183601 +0000 UTC m=+1245.686401231" lastFinishedPulling="2025-12-10 14:44:59.601235499 +0000 UTC m=+1249.170453129" observedRunningTime="2025-12-10 14:45:01.304032881 +0000 UTC m=+1250.873250531" watchObservedRunningTime="2025-12-10 14:45:01.315785647 +0000 UTC m=+1250.885003287" Dec 10 14:45:01 crc kubenswrapper[4847]: I1210 14:45:01.340283 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.064846859 podStartE2EDuration="6.340255684s" podCreationTimestamp="2025-12-10 14:44:55 +0000 UTC" firstStartedPulling="2025-12-10 14:44:56.343845803 +0000 UTC m=+1245.913063423" lastFinishedPulling="2025-12-10 14:44:59.619254618 +0000 UTC m=+1249.188472248" observedRunningTime="2025-12-10 14:45:01.331579928 +0000 UTC m=+1250.900797558" watchObservedRunningTime="2025-12-10 14:45:01.340255684 +0000 UTC m=+1250.909473314" Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.316362 4847 generic.go:334] "Generic (PLEG): container finished" podID="183c9599-f0db-4e64-9ad1-3c4b60bef02c" containerID="4b74eee694f848bebd50249943b03c3be4761073a25469c692597515fc7d0a3f" exitCode=0 Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.316570 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" event={"ID":"183c9599-f0db-4e64-9ad1-3c4b60bef02c","Type":"ContainerDied","Data":"4b74eee694f848bebd50249943b03c3be4761073a25469c692597515fc7d0a3f"} Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.320844 4847 generic.go:334] "Generic (PLEG): container finished" podID="ff251a94-9385-4610-a270-c9dea5c56272" containerID="af8d5747a7d81ff701f5b459dee8a83be0f14b7ce087be5bfb7ffe27ff7e60d5" exitCode=0 Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.320886 4847 generic.go:334] "Generic (PLEG): container finished" podID="ff251a94-9385-4610-a270-c9dea5c56272" containerID="5a715d0deacfa5eab8ca5886fad6232acb9fc56851e5b57afeb867b087731eb4" exitCode=143 Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.320883 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerDied","Data":"af8d5747a7d81ff701f5b459dee8a83be0f14b7ce087be5bfb7ffe27ff7e60d5"} Dec 10 14:45:02 crc kubenswrapper[4847]: I1210 14:45:02.320925 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerDied","Data":"5a715d0deacfa5eab8ca5886fad6232acb9fc56851e5b57afeb867b087731eb4"} Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.018310 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.170449 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle\") pod \"ff251a94-9385-4610-a270-c9dea5c56272\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.170805 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs\") pod \"ff251a94-9385-4610-a270-c9dea5c56272\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.170843 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29rdp\" (UniqueName: \"kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp\") pod \"ff251a94-9385-4610-a270-c9dea5c56272\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.171317 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data\") pod \"ff251a94-9385-4610-a270-c9dea5c56272\" (UID: \"ff251a94-9385-4610-a270-c9dea5c56272\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.172930 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs" (OuterVolumeSpecName: "logs") pod "ff251a94-9385-4610-a270-c9dea5c56272" (UID: "ff251a94-9385-4610-a270-c9dea5c56272"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.179674 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp" (OuterVolumeSpecName: "kube-api-access-29rdp") pod "ff251a94-9385-4610-a270-c9dea5c56272" (UID: "ff251a94-9385-4610-a270-c9dea5c56272"). InnerVolumeSpecName "kube-api-access-29rdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.205519 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff251a94-9385-4610-a270-c9dea5c56272" (UID: "ff251a94-9385-4610-a270-c9dea5c56272"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.212965 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data" (OuterVolumeSpecName: "config-data") pod "ff251a94-9385-4610-a270-c9dea5c56272" (UID: "ff251a94-9385-4610-a270-c9dea5c56272"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.274629 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.274672 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff251a94-9385-4610-a270-c9dea5c56272-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.274684 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29rdp\" (UniqueName: \"kubernetes.io/projected/ff251a94-9385-4610-a270-c9dea5c56272-kube-api-access-29rdp\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.274698 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff251a94-9385-4610-a270-c9dea5c56272-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.333319 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff251a94-9385-4610-a270-c9dea5c56272","Type":"ContainerDied","Data":"0eb9584970868baf062b89fb11040a0e7a6e26dba68009b246d351963e29d4a9"} Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.335563 4847 scope.go:117] "RemoveContainer" containerID="af8d5747a7d81ff701f5b459dee8a83be0f14b7ce087be5bfb7ffe27ff7e60d5" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.333528 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.357129 4847 scope.go:117] "RemoveContainer" containerID="5a715d0deacfa5eab8ca5886fad6232acb9fc56851e5b57afeb867b087731eb4" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.400628 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.419318 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.458868 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:03 crc kubenswrapper[4847]: E1210 14:45:03.459477 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-log" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.459497 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-log" Dec 10 14:45:03 crc kubenswrapper[4847]: E1210 14:45:03.459515 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-metadata" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.459522 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-metadata" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.459862 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-metadata" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.459886 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff251a94-9385-4610-a270-c9dea5c56272" containerName="nova-metadata-log" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.461056 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.463701 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.463921 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.491466 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.588867 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.589210 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.589239 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.589322 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vhzs\" (UniqueName: \"kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.589461 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.690860 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.690935 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.690970 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.691000 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.691068 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vhzs\" (UniqueName: \"kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.691451 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.698677 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.709126 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.711199 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.711831 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vhzs\" (UniqueName: \"kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs\") pod \"nova-metadata-0\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.742583 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.786659 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.896246 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume\") pod \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.896325 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lzc6\" (UniqueName: \"kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6\") pod \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.896407 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume\") pod \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\" (UID: \"183c9599-f0db-4e64-9ad1-3c4b60bef02c\") " Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.896779 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume" (OuterVolumeSpecName: "config-volume") pod "183c9599-f0db-4e64-9ad1-3c4b60bef02c" (UID: "183c9599-f0db-4e64-9ad1-3c4b60bef02c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.898112 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/183c9599-f0db-4e64-9ad1-3c4b60bef02c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.901545 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "183c9599-f0db-4e64-9ad1-3c4b60bef02c" (UID: "183c9599-f0db-4e64-9ad1-3c4b60bef02c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:03 crc kubenswrapper[4847]: I1210 14:45:03.901744 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6" (OuterVolumeSpecName: "kube-api-access-2lzc6") pod "183c9599-f0db-4e64-9ad1-3c4b60bef02c" (UID: "183c9599-f0db-4e64-9ad1-3c4b60bef02c"). InnerVolumeSpecName "kube-api-access-2lzc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:03.999829 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lzc6\" (UniqueName: \"kubernetes.io/projected/183c9599-f0db-4e64-9ad1-3c4b60bef02c-kube-api-access-2lzc6\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.000150 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/183c9599-f0db-4e64-9ad1-3c4b60bef02c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.240692 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:04 crc kubenswrapper[4847]: W1210 14:45:04.247554 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c85f3a7_58aa_401b_85cf_eac8c9c61832.slice/crio-630b2a13a30f3ed9d9552d80a2667bb5d983f064477ab5fd1ef8353cf864dfa7 WatchSource:0}: Error finding container 630b2a13a30f3ed9d9552d80a2667bb5d983f064477ab5fd1ef8353cf864dfa7: Status 404 returned error can't find the container with id 630b2a13a30f3ed9d9552d80a2667bb5d983f064477ab5fd1ef8353cf864dfa7 Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.354743 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerStarted","Data":"630b2a13a30f3ed9d9552d80a2667bb5d983f064477ab5fd1ef8353cf864dfa7"} Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.360616 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" event={"ID":"183c9599-f0db-4e64-9ad1-3c4b60bef02c","Type":"ContainerDied","Data":"2e45b9de4c72e243ea7edc7e2fb07620f6ba570f157c904f36bba4ce9f1fa325"} Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.360657 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e45b9de4c72e243ea7edc7e2fb07620f6ba570f157c904f36bba4ce9f1fa325" Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.360729 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7" Dec 10 14:45:04 crc kubenswrapper[4847]: I1210 14:45:04.776574 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff251a94-9385-4610-a270-c9dea5c56272" path="/var/lib/kubelet/pods/ff251a94-9385-4610-a270-c9dea5c56272/volumes" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.381450 4847 generic.go:334] "Generic (PLEG): container finished" podID="5dd1e94d-67a9-443c-9870-81a88bf56445" containerID="cd7aebf5817a579950b349a82d3860e10ad53f542e61f2fd453d39385993045d" exitCode=0 Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.381545 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zdcv2" event={"ID":"5dd1e94d-67a9-443c-9870-81a88bf56445","Type":"ContainerDied","Data":"cd7aebf5817a579950b349a82d3860e10ad53f542e61f2fd453d39385993045d"} Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.387507 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerStarted","Data":"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d"} Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.388450 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerStarted","Data":"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8"} Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.429659 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.429623371 podStartE2EDuration="2.429623371s" podCreationTimestamp="2025-12-10 14:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:05.422910376 +0000 UTC m=+1254.992128006" watchObservedRunningTime="2025-12-10 14:45:05.429623371 +0000 UTC m=+1254.998841001" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.471073 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.471146 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.484704 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.514645 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.723927 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.802962 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:45:05 crc kubenswrapper[4847]: I1210 14:45:05.803551 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="dnsmasq-dns" containerID="cri-o://a710b931fedbf0a85ddad2347e83365594a0b23b9583983cc7f44422aaaac81e" gracePeriod=10 Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.414030 4847 generic.go:334] "Generic (PLEG): container finished" podID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerID="a710b931fedbf0a85ddad2347e83365594a0b23b9583983cc7f44422aaaac81e" exitCode=0 Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.414313 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" event={"ID":"159c593e-de1b-4dfe-b846-d0da68a0881a","Type":"ContainerDied","Data":"a710b931fedbf0a85ddad2347e83365594a0b23b9583983cc7f44422aaaac81e"} Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.414562 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" event={"ID":"159c593e-de1b-4dfe-b846-d0da68a0881a","Type":"ContainerDied","Data":"fe148ffc66510b6b23f27bcf718f797926843935590a6947f969e258b90699d2"} Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.414578 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe148ffc66510b6b23f27bcf718f797926843935590a6947f969e258b90699d2" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.444028 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.450590 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.480807 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.480966 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.481739 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.481788 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npmt9\" (UniqueName: \"kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.481891 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.481928 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.503140 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9" (OuterVolumeSpecName: "kube-api-access-npmt9") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "kube-api-access-npmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.549333 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config" (OuterVolumeSpecName: "config") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.557460 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.557751 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.590858 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.591263 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") pod \"159c593e-de1b-4dfe-b846-d0da68a0881a\" (UID: \"159c593e-de1b-4dfe-b846-d0da68a0881a\") " Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.591858 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.591888 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npmt9\" (UniqueName: \"kubernetes.io/projected/159c593e-de1b-4dfe-b846-d0da68a0881a-kube-api-access-npmt9\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: W1210 14:45:06.592000 4847 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/159c593e-de1b-4dfe-b846-d0da68a0881a/volumes/kubernetes.io~configmap/dns-svc Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.592057 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.600315 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.600392 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.615059 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "159c593e-de1b-4dfe-b846-d0da68a0881a" (UID: "159c593e-de1b-4dfe-b846-d0da68a0881a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.694454 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.694503 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.694517 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.694529 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/159c593e-de1b-4dfe-b846-d0da68a0881a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:06 crc kubenswrapper[4847]: I1210 14:45:06.853289 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.009826 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data\") pod \"5dd1e94d-67a9-443c-9870-81a88bf56445\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.010514 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle\") pod \"5dd1e94d-67a9-443c-9870-81a88bf56445\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.010632 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hsbm\" (UniqueName: \"kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm\") pod \"5dd1e94d-67a9-443c-9870-81a88bf56445\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.010688 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts\") pod \"5dd1e94d-67a9-443c-9870-81a88bf56445\" (UID: \"5dd1e94d-67a9-443c-9870-81a88bf56445\") " Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.026510 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts" (OuterVolumeSpecName: "scripts") pod "5dd1e94d-67a9-443c-9870-81a88bf56445" (UID: "5dd1e94d-67a9-443c-9870-81a88bf56445"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.042892 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm" (OuterVolumeSpecName: "kube-api-access-5hsbm") pod "5dd1e94d-67a9-443c-9870-81a88bf56445" (UID: "5dd1e94d-67a9-443c-9870-81a88bf56445"). InnerVolumeSpecName "kube-api-access-5hsbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.113313 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.113348 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hsbm\" (UniqueName: \"kubernetes.io/projected/5dd1e94d-67a9-443c-9870-81a88bf56445-kube-api-access-5hsbm\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.117099 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5dd1e94d-67a9-443c-9870-81a88bf56445" (UID: "5dd1e94d-67a9-443c-9870-81a88bf56445"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.117296 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data" (OuterVolumeSpecName: "config-data") pod "5dd1e94d-67a9-443c-9870-81a88bf56445" (UID: "5dd1e94d-67a9-443c-9870-81a88bf56445"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.214879 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.214915 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dd1e94d-67a9-443c-9870-81a88bf56445-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.429535 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-q4q79" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.431106 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zdcv2" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.435498 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zdcv2" event={"ID":"5dd1e94d-67a9-443c-9870-81a88bf56445","Type":"ContainerDied","Data":"ac82dc16abaae323f94a244ad6b13ae4cd8d7bc5c4acc146a13015c23cd17629"} Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.435550 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac82dc16abaae323f94a244ad6b13ae4cd8d7bc5c4acc146a13015c23cd17629" Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.469877 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.477998 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-q4q79"] Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.544447 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.544660 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-log" containerID="cri-o://608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32" gracePeriod=30 Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.544887 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-api" containerID="cri-o://102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449" gracePeriod=30 Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.575112 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.575375 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-log" containerID="cri-o://2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" gracePeriod=30 Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.575904 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-metadata" containerID="cri-o://79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" gracePeriod=30 Dec 10 14:45:07 crc kubenswrapper[4847]: I1210 14:45:07.925961 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.149347 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.337548 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs\") pod \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.337662 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs\") pod \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.337773 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data\") pod \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.337802 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle\") pod \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.337950 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vhzs\" (UniqueName: \"kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs\") pod \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\" (UID: \"6c85f3a7-58aa-401b-85cf-eac8c9c61832\") " Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.339080 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs" (OuterVolumeSpecName: "logs") pod "6c85f3a7-58aa-401b-85cf-eac8c9c61832" (UID: "6c85f3a7-58aa-401b-85cf-eac8c9c61832"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.350005 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs" (OuterVolumeSpecName: "kube-api-access-8vhzs") pod "6c85f3a7-58aa-401b-85cf-eac8c9c61832" (UID: "6c85f3a7-58aa-401b-85cf-eac8c9c61832"). InnerVolumeSpecName "kube-api-access-8vhzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.370043 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c85f3a7-58aa-401b-85cf-eac8c9c61832" (UID: "6c85f3a7-58aa-401b-85cf-eac8c9c61832"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.376567 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data" (OuterVolumeSpecName: "config-data") pod "6c85f3a7-58aa-401b-85cf-eac8c9c61832" (UID: "6c85f3a7-58aa-401b-85cf-eac8c9c61832"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.400032 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6c85f3a7-58aa-401b-85cf-eac8c9c61832" (UID: "6c85f3a7-58aa-401b-85cf-eac8c9c61832"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.439820 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vhzs\" (UniqueName: \"kubernetes.io/projected/6c85f3a7-58aa-401b-85cf-eac8c9c61832-kube-api-access-8vhzs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.440262 4847 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.440278 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c85f3a7-58aa-401b-85cf-eac8c9c61832-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.440291 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.440305 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c85f3a7-58aa-401b-85cf-eac8c9c61832-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.445917 4847 generic.go:334] "Generic (PLEG): container finished" podID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerID="608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32" exitCode=143 Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.446003 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerDied","Data":"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32"} Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450293 4847 generic.go:334] "Generic (PLEG): container finished" podID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerID="79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" exitCode=0 Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450336 4847 generic.go:334] "Generic (PLEG): container finished" podID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerID="2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" exitCode=143 Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450404 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerDied","Data":"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d"} Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450441 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerDied","Data":"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8"} Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450457 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6c85f3a7-58aa-401b-85cf-eac8c9c61832","Type":"ContainerDied","Data":"630b2a13a30f3ed9d9552d80a2667bb5d983f064477ab5fd1ef8353cf864dfa7"} Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450476 4847 scope.go:117] "RemoveContainer" containerID="79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.450648 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.457224 4847 generic.go:334] "Generic (PLEG): container finished" podID="f0ae39e9-b424-481b-a316-403a59eb4837" containerID="68a84f74c916db4f5a2cbd1b3d639386df8b56b0b4509c4be4dde8a49eba44cf" exitCode=0 Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.457430 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerName="nova-scheduler-scheduler" containerID="cri-o://1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" gracePeriod=30 Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.457804 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" event={"ID":"f0ae39e9-b424-481b-a316-403a59eb4837","Type":"ContainerDied","Data":"68a84f74c916db4f5a2cbd1b3d639386df8b56b0b4509c4be4dde8a49eba44cf"} Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.582083 4847 scope.go:117] "RemoveContainer" containerID="2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.590981 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.599167 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.621079 4847 scope.go:117] "RemoveContainer" containerID="79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.622165 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d\": container with ID starting with 79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d not found: ID does not exist" containerID="79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.622250 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d"} err="failed to get container status \"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d\": rpc error: code = NotFound desc = could not find container \"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d\": container with ID starting with 79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d not found: ID does not exist" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.622332 4847 scope.go:117] "RemoveContainer" containerID="2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.631178 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8\": container with ID starting with 2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8 not found: ID does not exist" containerID="2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.631445 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8"} err="failed to get container status \"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8\": rpc error: code = NotFound desc = could not find container \"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8\": container with ID starting with 2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8 not found: ID does not exist" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.631527 4847 scope.go:117] "RemoveContainer" containerID="79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.633754 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d"} err="failed to get container status \"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d\": rpc error: code = NotFound desc = could not find container \"79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d\": container with ID starting with 79ba8ff52a00af6ce0905221aed4aab8dc38b2912827f2b0aaa640f523b8928d not found: ID does not exist" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.633858 4847 scope.go:117] "RemoveContainer" containerID="2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.634773 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8"} err="failed to get container status \"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8\": rpc error: code = NotFound desc = could not find container \"2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8\": container with ID starting with 2d06f6a98d7172b14c98c0acd63d7059af9d972a413e1594f29e6a7fbc963ad8 not found: ID does not exist" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.636384 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637050 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="init" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637076 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="init" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637087 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-log" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637095 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-log" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637112 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="dnsmasq-dns" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637122 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="dnsmasq-dns" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637150 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dd1e94d-67a9-443c-9870-81a88bf56445" containerName="nova-manage" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637163 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dd1e94d-67a9-443c-9870-81a88bf56445" containerName="nova-manage" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637174 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="183c9599-f0db-4e64-9ad1-3c4b60bef02c" containerName="collect-profiles" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637183 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="183c9599-f0db-4e64-9ad1-3c4b60bef02c" containerName="collect-profiles" Dec 10 14:45:08 crc kubenswrapper[4847]: E1210 14:45:08.637202 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-metadata" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637211 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-metadata" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637482 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" containerName="dnsmasq-dns" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637506 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-log" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637522 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dd1e94d-67a9-443c-9870-81a88bf56445" containerName="nova-manage" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637535 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" containerName="nova-metadata-metadata" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.637544 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="183c9599-f0db-4e64-9ad1-3c4b60bef02c" containerName="collect-profiles" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.638968 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.647106 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.648094 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.673919 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.749207 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.749303 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.749369 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.749394 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.749485 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6g9q\" (UniqueName: \"kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.773311 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="159c593e-de1b-4dfe-b846-d0da68a0881a" path="/var/lib/kubelet/pods/159c593e-de1b-4dfe-b846-d0da68a0881a/volumes" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.774104 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c85f3a7-58aa-401b-85cf-eac8c9c61832" path="/var/lib/kubelet/pods/6c85f3a7-58aa-401b-85cf-eac8c9c61832/volumes" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.851645 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.851801 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.851839 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.851871 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.851926 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6g9q\" (UniqueName: \"kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.853752 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.858116 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.858153 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.861148 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.875835 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6g9q\" (UniqueName: \"kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q\") pod \"nova-metadata-0\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " pod="openstack/nova-metadata-0" Dec 10 14:45:08 crc kubenswrapper[4847]: I1210 14:45:08.984180 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:09 crc kubenswrapper[4847]: I1210 14:45:09.505282 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:09 crc kubenswrapper[4847]: I1210 14:45:09.566650 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 14:45:09 crc kubenswrapper[4847]: I1210 14:45:09.932145 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.075320 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data\") pod \"f0ae39e9-b424-481b-a316-403a59eb4837\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.075367 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle\") pod \"f0ae39e9-b424-481b-a316-403a59eb4837\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.075428 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zflb\" (UniqueName: \"kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb\") pod \"f0ae39e9-b424-481b-a316-403a59eb4837\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.075495 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts\") pod \"f0ae39e9-b424-481b-a316-403a59eb4837\" (UID: \"f0ae39e9-b424-481b-a316-403a59eb4837\") " Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.080862 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts" (OuterVolumeSpecName: "scripts") pod "f0ae39e9-b424-481b-a316-403a59eb4837" (UID: "f0ae39e9-b424-481b-a316-403a59eb4837"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.080961 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb" (OuterVolumeSpecName: "kube-api-access-4zflb") pod "f0ae39e9-b424-481b-a316-403a59eb4837" (UID: "f0ae39e9-b424-481b-a316-403a59eb4837"). InnerVolumeSpecName "kube-api-access-4zflb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.104301 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0ae39e9-b424-481b-a316-403a59eb4837" (UID: "f0ae39e9-b424-481b-a316-403a59eb4837"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.104704 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data" (OuterVolumeSpecName: "config-data") pod "f0ae39e9-b424-481b-a316-403a59eb4837" (UID: "f0ae39e9-b424-481b-a316-403a59eb4837"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.190873 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.191322 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.191342 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zflb\" (UniqueName: \"kubernetes.io/projected/f0ae39e9-b424-481b-a316-403a59eb4837-kube-api-access-4zflb\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.191356 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0ae39e9-b424-481b-a316-403a59eb4837-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.479907 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.479929 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8bgjb" event={"ID":"f0ae39e9-b424-481b-a316-403a59eb4837","Type":"ContainerDied","Data":"15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb"} Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.480047 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15dec20b40e3fdb88deee5949044aa161716ebadbc919c796c5d11c2cb4bdacb" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.483817 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerStarted","Data":"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733"} Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.483861 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerStarted","Data":"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231"} Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.483875 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerStarted","Data":"cceade013397a1492a8ddcc3b0250468f69711dd8f90ce0e0e33e280caca8f41"} Dec 10 14:45:10 crc kubenswrapper[4847]: E1210 14:45:10.487412 4847 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 14:45:10 crc kubenswrapper[4847]: E1210 14:45:10.490749 4847 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 14:45:10 crc kubenswrapper[4847]: E1210 14:45:10.493529 4847 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 14:45:10 crc kubenswrapper[4847]: E1210 14:45:10.493564 4847 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerName="nova-scheduler-scheduler" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.512068 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.512051672 podStartE2EDuration="2.512051672s" podCreationTimestamp="2025-12-10 14:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:10.506878628 +0000 UTC m=+1260.076096248" watchObservedRunningTime="2025-12-10 14:45:10.512051672 +0000 UTC m=+1260.081269292" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.586628 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 14:45:10 crc kubenswrapper[4847]: E1210 14:45:10.587094 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0ae39e9-b424-481b-a316-403a59eb4837" containerName="nova-cell1-conductor-db-sync" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.587112 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0ae39e9-b424-481b-a316-403a59eb4837" containerName="nova-cell1-conductor-db-sync" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.587286 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0ae39e9-b424-481b-a316-403a59eb4837" containerName="nova-cell1-conductor-db-sync" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.588194 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.596425 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.604488 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.702533 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.702591 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb87q\" (UniqueName: \"kubernetes.io/projected/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-kube-api-access-mb87q\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.702659 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.806487 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.806566 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb87q\" (UniqueName: \"kubernetes.io/projected/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-kube-api-access-mb87q\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.806809 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.809265 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.811473 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.820346 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.823392 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb87q\" (UniqueName: \"kubernetes.io/projected/1598cbc7-d655-494f-a2b1-ddbd4db56aa5-kube-api-access-mb87q\") pod \"nova-cell1-conductor-0\" (UID: \"1598cbc7-d655-494f-a2b1-ddbd4db56aa5\") " pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:10 crc kubenswrapper[4847]: I1210 14:45:10.912258 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:11 crc kubenswrapper[4847]: W1210 14:45:11.423044 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1598cbc7_d655_494f_a2b1_ddbd4db56aa5.slice/crio-0402205590e6bea5c11d9c2262fcaa001247275531f8fecdf280496571cb36b8 WatchSource:0}: Error finding container 0402205590e6bea5c11d9c2262fcaa001247275531f8fecdf280496571cb36b8: Status 404 returned error can't find the container with id 0402205590e6bea5c11d9c2262fcaa001247275531f8fecdf280496571cb36b8 Dec 10 14:45:11 crc kubenswrapper[4847]: I1210 14:45:11.430324 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 14:45:11 crc kubenswrapper[4847]: I1210 14:45:11.496194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1598cbc7-d655-494f-a2b1-ddbd4db56aa5","Type":"ContainerStarted","Data":"0402205590e6bea5c11d9c2262fcaa001247275531f8fecdf280496571cb36b8"} Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.535513 4847 generic.go:334] "Generic (PLEG): container finished" podID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerID="1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" exitCode=0 Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.536082 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23c0a0d3-8946-4120-b757-d096b31f67c8","Type":"ContainerDied","Data":"1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae"} Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.554672 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1598cbc7-d655-494f-a2b1-ddbd4db56aa5","Type":"ContainerStarted","Data":"2655b0abbbd490bc3ae69ead49dafa50098c783e1742b62b6c285ecbe4f76dd7"} Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.555356 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.576559 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.5765419510000003 podStartE2EDuration="2.576541951s" podCreationTimestamp="2025-12-10 14:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:12.572381233 +0000 UTC m=+1262.141598893" watchObservedRunningTime="2025-12-10 14:45:12.576541951 +0000 UTC m=+1262.145759581" Dec 10 14:45:12 crc kubenswrapper[4847]: I1210 14:45:12.873207 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.049427 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xrvq\" (UniqueName: \"kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq\") pod \"23c0a0d3-8946-4120-b757-d096b31f67c8\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.049683 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle\") pod \"23c0a0d3-8946-4120-b757-d096b31f67c8\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.049758 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data\") pod \"23c0a0d3-8946-4120-b757-d096b31f67c8\" (UID: \"23c0a0d3-8946-4120-b757-d096b31f67c8\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.061216 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq" (OuterVolumeSpecName: "kube-api-access-9xrvq") pod "23c0a0d3-8946-4120-b757-d096b31f67c8" (UID: "23c0a0d3-8946-4120-b757-d096b31f67c8"). InnerVolumeSpecName "kube-api-access-9xrvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.084040 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data" (OuterVolumeSpecName: "config-data") pod "23c0a0d3-8946-4120-b757-d096b31f67c8" (UID: "23c0a0d3-8946-4120-b757-d096b31f67c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.103885 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23c0a0d3-8946-4120-b757-d096b31f67c8" (UID: "23c0a0d3-8946-4120-b757-d096b31f67c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.152512 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.152547 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c0a0d3-8946-4120-b757-d096b31f67c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.152557 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xrvq\" (UniqueName: \"kubernetes.io/projected/23c0a0d3-8946-4120-b757-d096b31f67c8-kube-api-access-9xrvq\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.353697 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.354922 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data\") pod \"61e52874-0fa5-4688-ac1e-33fae92b179f\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.354986 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp7ct\" (UniqueName: \"kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct\") pod \"61e52874-0fa5-4688-ac1e-33fae92b179f\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.355029 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle\") pod \"61e52874-0fa5-4688-ac1e-33fae92b179f\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.364655 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct" (OuterVolumeSpecName: "kube-api-access-mp7ct") pod "61e52874-0fa5-4688-ac1e-33fae92b179f" (UID: "61e52874-0fa5-4688-ac1e-33fae92b179f"). InnerVolumeSpecName "kube-api-access-mp7ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.400376 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data" (OuterVolumeSpecName: "config-data") pod "61e52874-0fa5-4688-ac1e-33fae92b179f" (UID: "61e52874-0fa5-4688-ac1e-33fae92b179f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.427761 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61e52874-0fa5-4688-ac1e-33fae92b179f" (UID: "61e52874-0fa5-4688-ac1e-33fae92b179f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.456290 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs\") pod \"61e52874-0fa5-4688-ac1e-33fae92b179f\" (UID: \"61e52874-0fa5-4688-ac1e-33fae92b179f\") " Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.456568 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.456588 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp7ct\" (UniqueName: \"kubernetes.io/projected/61e52874-0fa5-4688-ac1e-33fae92b179f-kube-api-access-mp7ct\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.456598 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61e52874-0fa5-4688-ac1e-33fae92b179f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.457003 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs" (OuterVolumeSpecName: "logs") pod "61e52874-0fa5-4688-ac1e-33fae92b179f" (UID: "61e52874-0fa5-4688-ac1e-33fae92b179f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.558064 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61e52874-0fa5-4688-ac1e-33fae92b179f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.566580 4847 generic.go:334] "Generic (PLEG): container finished" podID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerID="102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449" exitCode=0 Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.566660 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerDied","Data":"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449"} Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.566693 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"61e52874-0fa5-4688-ac1e-33fae92b179f","Type":"ContainerDied","Data":"f10a7ce47b2c1973402880825983040ef7ab2431f121a8271ea941566cc95d0c"} Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.566730 4847 scope.go:117] "RemoveContainer" containerID="102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.566873 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.589667 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.589754 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"23c0a0d3-8946-4120-b757-d096b31f67c8","Type":"ContainerDied","Data":"6ba6c4b1a827271949b0e4e4c3af499fe38f248d06e429c47bc8578f11ea7d97"} Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.626903 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.633309 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.636984 4847 scope.go:117] "RemoveContainer" containerID="608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.680788 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.697940 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.710874 4847 scope.go:117] "RemoveContainer" containerID="102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449" Dec 10 14:45:13 crc kubenswrapper[4847]: E1210 14:45:13.713748 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449\": container with ID starting with 102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449 not found: ID does not exist" containerID="102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.713778 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449"} err="failed to get container status \"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449\": rpc error: code = NotFound desc = could not find container \"102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449\": container with ID starting with 102d4fd1ce49e1ae6461c27394b3c57563c5ecde92c9a584d38579294a68f449 not found: ID does not exist" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.713799 4847 scope.go:117] "RemoveContainer" containerID="608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.713941 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: E1210 14:45:13.714336 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-log" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714356 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-log" Dec 10 14:45:13 crc kubenswrapper[4847]: E1210 14:45:13.714378 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerName="nova-scheduler-scheduler" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714386 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerName="nova-scheduler-scheduler" Dec 10 14:45:13 crc kubenswrapper[4847]: E1210 14:45:13.714400 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-api" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714407 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-api" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714612 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-log" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714631 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" containerName="nova-scheduler-scheduler" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.714642 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" containerName="nova-api-api" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.715587 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: E1210 14:45:13.716259 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32\": container with ID starting with 608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32 not found: ID does not exist" containerID="608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.716340 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32"} err="failed to get container status \"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32\": rpc error: code = NotFound desc = could not find container \"608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32\": container with ID starting with 608615f959ec5e8ac41060ef21486c539a5da985e5eb655e4e74ee847a7fec32 not found: ID does not exist" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.716383 4847 scope.go:117] "RemoveContainer" containerID="1bad48f8c61b95c9644d514d6d22af1cb0baf1daa06561c863adb6d72bf23dae" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.720229 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.736844 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.738070 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.741353 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.759133 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776673 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whjzr\" (UniqueName: \"kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776742 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776765 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776788 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776833 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776848 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.776884 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4fdf\" (UniqueName: \"kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.797783 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.879834 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whjzr\" (UniqueName: \"kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.879926 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.879965 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.880008 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.880092 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.880150 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.880623 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.881025 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4fdf\" (UniqueName: \"kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.888414 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.888474 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.899870 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.900823 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whjzr\" (UniqueName: \"kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.906485 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " pod="openstack/nova-api-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.908387 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4fdf\" (UniqueName: \"kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf\") pod \"nova-scheduler-0\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.985066 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 14:45:13 crc kubenswrapper[4847]: I1210 14:45:13.985113 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.053766 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.102668 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.613222 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.704579 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.705047 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4d586df7-072d-4e8f-b2c2-250c31c20a29" containerName="kube-state-metrics" containerID="cri-o://45fbd8f086bedf06905f3c444934fb69237df6a3ad9d5d1758c7eea701f28abb" gracePeriod=30 Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.719378 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:14 crc kubenswrapper[4847]: W1210 14:45:14.723850 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d2d1859_c765_4069_9ad6_8c919c5d2783.slice/crio-d4fb498287540bad39aa853cb6d7cba4c6d482c43dc451dc665a4f234306b646 WatchSource:0}: Error finding container d4fb498287540bad39aa853cb6d7cba4c6d482c43dc451dc665a4f234306b646: Status 404 returned error can't find the container with id d4fb498287540bad39aa853cb6d7cba4c6d482c43dc451dc665a4f234306b646 Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.775470 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23c0a0d3-8946-4120-b757-d096b31f67c8" path="/var/lib/kubelet/pods/23c0a0d3-8946-4120-b757-d096b31f67c8/volumes" Dec 10 14:45:14 crc kubenswrapper[4847]: I1210 14:45:14.778515 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61e52874-0fa5-4688-ac1e-33fae92b179f" path="/var/lib/kubelet/pods/61e52874-0fa5-4688-ac1e-33fae92b179f/volumes" Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.617152 4847 generic.go:334] "Generic (PLEG): container finished" podID="4d586df7-072d-4e8f-b2c2-250c31c20a29" containerID="45fbd8f086bedf06905f3c444934fb69237df6a3ad9d5d1758c7eea701f28abb" exitCode=2 Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.617517 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d586df7-072d-4e8f-b2c2-250c31c20a29","Type":"ContainerDied","Data":"45fbd8f086bedf06905f3c444934fb69237df6a3ad9d5d1758c7eea701f28abb"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.619012 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d2d1859-c765-4069-9ad6-8c919c5d2783","Type":"ContainerStarted","Data":"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.619051 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d2d1859-c765-4069-9ad6-8c919c5d2783","Type":"ContainerStarted","Data":"d4fb498287540bad39aa853cb6d7cba4c6d482c43dc451dc665a4f234306b646"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.622090 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerStarted","Data":"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.622121 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerStarted","Data":"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.622130 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerStarted","Data":"cf05691bf39ddb00d00bc1bb19eba0c10edad39ca88420912e8acffe5fa9368c"} Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.649438 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.649416603 podStartE2EDuration="2.649416603s" podCreationTimestamp="2025-12-10 14:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:15.645264385 +0000 UTC m=+1265.214482025" watchObservedRunningTime="2025-12-10 14:45:15.649416603 +0000 UTC m=+1265.218634233" Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.662888 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.662865963 podStartE2EDuration="2.662865963s" podCreationTimestamp="2025-12-10 14:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:15.660458801 +0000 UTC m=+1265.229676431" watchObservedRunningTime="2025-12-10 14:45:15.662865963 +0000 UTC m=+1265.232083593" Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.713283 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.822244 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnmg5\" (UniqueName: \"kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5\") pod \"4d586df7-072d-4e8f-b2c2-250c31c20a29\" (UID: \"4d586df7-072d-4e8f-b2c2-250c31c20a29\") " Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.829371 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5" (OuterVolumeSpecName: "kube-api-access-cnmg5") pod "4d586df7-072d-4e8f-b2c2-250c31c20a29" (UID: "4d586df7-072d-4e8f-b2c2-250c31c20a29"). InnerVolumeSpecName "kube-api-access-cnmg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:15 crc kubenswrapper[4847]: I1210 14:45:15.924136 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnmg5\" (UniqueName: \"kubernetes.io/projected/4d586df7-072d-4e8f-b2c2-250c31c20a29-kube-api-access-cnmg5\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.635232 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d586df7-072d-4e8f-b2c2-250c31c20a29","Type":"ContainerDied","Data":"f5e244c99713928a7b216d10294ace5a596215d47fa35ad025f6306dd78a8367"} Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.635289 4847 scope.go:117] "RemoveContainer" containerID="45fbd8f086bedf06905f3c444934fb69237df6a3ad9d5d1758c7eea701f28abb" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.635452 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.686097 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.700290 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.739615 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:16 crc kubenswrapper[4847]: E1210 14:45:16.740246 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d586df7-072d-4e8f-b2c2-250c31c20a29" containerName="kube-state-metrics" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.740268 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d586df7-072d-4e8f-b2c2-250c31c20a29" containerName="kube-state-metrics" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.740522 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d586df7-072d-4e8f-b2c2-250c31c20a29" containerName="kube-state-metrics" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.741432 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.747238 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.747340 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.756438 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.800478 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d586df7-072d-4e8f-b2c2-250c31c20a29" path="/var/lib/kubelet/pods/4d586df7-072d-4e8f-b2c2-250c31c20a29/volumes" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.843804 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5v58\" (UniqueName: \"kubernetes.io/projected/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-api-access-b5v58\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.843967 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.844001 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.844138 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.945525 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.945572 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.945754 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.945779 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5v58\" (UniqueName: \"kubernetes.io/projected/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-api-access-b5v58\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.956317 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.957221 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.981307 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:16 crc kubenswrapper[4847]: I1210 14:45:16.984589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5v58\" (UniqueName: \"kubernetes.io/projected/6511a21a-53cd-4ecc-a550-cc4a3ef72c8c-kube-api-access-b5v58\") pod \"kube-state-metrics-0\" (UID: \"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c\") " pod="openstack/kube-state-metrics-0" Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.062850 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.248644 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.249414 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-central-agent" containerID="cri-o://97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1" gracePeriod=30 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.250074 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="proxy-httpd" containerID="cri-o://e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5" gracePeriod=30 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.250138 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="sg-core" containerID="cri-o://f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9" gracePeriod=30 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.250184 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-notification-agent" containerID="cri-o://9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec" gracePeriod=30 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.598565 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 14:45:17 crc kubenswrapper[4847]: W1210 14:45:17.602416 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6511a21a_53cd_4ecc_a550_cc4a3ef72c8c.slice/crio-6b9087a4e936e0d1152d3a74be0bff42a2a94a72e3b9bf82983abbaf9441918e WatchSource:0}: Error finding container 6b9087a4e936e0d1152d3a74be0bff42a2a94a72e3b9bf82983abbaf9441918e: Status 404 returned error can't find the container with id 6b9087a4e936e0d1152d3a74be0bff42a2a94a72e3b9bf82983abbaf9441918e Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.649900 4847 generic.go:334] "Generic (PLEG): container finished" podID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerID="e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5" exitCode=0 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.649937 4847 generic.go:334] "Generic (PLEG): container finished" podID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerID="f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9" exitCode=2 Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.649972 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerDied","Data":"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5"} Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.650031 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerDied","Data":"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9"} Dec 10 14:45:17 crc kubenswrapper[4847]: I1210 14:45:17.654449 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c","Type":"ContainerStarted","Data":"6b9087a4e936e0d1152d3a74be0bff42a2a94a72e3b9bf82983abbaf9441918e"} Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.665119 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6511a21a-53cd-4ecc-a550-cc4a3ef72c8c","Type":"ContainerStarted","Data":"921ad0de06b254852fbf39acff41d57ffd8954191cfe2239bf7ad12d900174f4"} Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.665489 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.668503 4847 generic.go:334] "Generic (PLEG): container finished" podID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerID="97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1" exitCode=0 Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.668547 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerDied","Data":"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1"} Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.684582 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.3300315830000002 podStartE2EDuration="2.684564473s" podCreationTimestamp="2025-12-10 14:45:16 +0000 UTC" firstStartedPulling="2025-12-10 14:45:17.609041762 +0000 UTC m=+1267.178259392" lastFinishedPulling="2025-12-10 14:45:17.963574652 +0000 UTC m=+1267.532792282" observedRunningTime="2025-12-10 14:45:18.679890451 +0000 UTC m=+1268.249108081" watchObservedRunningTime="2025-12-10 14:45:18.684564473 +0000 UTC m=+1268.253782103" Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.985119 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 14:45:18 crc kubenswrapper[4847]: I1210 14:45:18.985167 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.103831 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.489353 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.606973 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607062 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607102 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607194 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607218 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7dqp\" (UniqueName: \"kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607386 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.607412 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml\") pod \"c74ba68c-4c24-488f-a863-e35cece9f1d9\" (UID: \"c74ba68c-4c24-488f-a863-e35cece9f1d9\") " Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.609244 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.612373 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.617851 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts" (OuterVolumeSpecName: "scripts") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.617945 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp" (OuterVolumeSpecName: "kube-api-access-v7dqp") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "kube-api-access-v7dqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.653351 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.679433 4847 generic.go:334] "Generic (PLEG): container finished" podID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerID="9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec" exitCode=0 Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.680290 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.680795 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerDied","Data":"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec"} Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.680821 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c74ba68c-4c24-488f-a863-e35cece9f1d9","Type":"ContainerDied","Data":"eb6f05822d47bbe0a768b5e175dd7ab7a0405792fd73958aef1a0ae2487b58b9"} Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.680836 4847 scope.go:117] "RemoveContainer" containerID="e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.692228 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710033 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710064 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710074 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710084 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710092 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c74ba68c-4c24-488f-a863-e35cece9f1d9-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.710102 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7dqp\" (UniqueName: \"kubernetes.io/projected/c74ba68c-4c24-488f-a863-e35cece9f1d9-kube-api-access-v7dqp\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.746603 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data" (OuterVolumeSpecName: "config-data") pod "c74ba68c-4c24-488f-a863-e35cece9f1d9" (UID: "c74ba68c-4c24-488f-a863-e35cece9f1d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:19 crc kubenswrapper[4847]: I1210 14:45:19.812160 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74ba68c-4c24-488f-a863-e35cece9f1d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:19.999890 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:19.999940 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.023642 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.023912 4847 scope.go:117] "RemoveContainer" containerID="f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.032746 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.099965 4847 scope.go:117] "RemoveContainer" containerID="9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.148797 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.149238 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="proxy-httpd" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149256 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="proxy-httpd" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.149273 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="sg-core" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149279 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="sg-core" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.149293 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-central-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149302 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-central-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.149314 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-notification-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149320 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-notification-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149489 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="sg-core" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149511 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-central-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149517 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="ceilometer-notification-agent" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.149524 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" containerName="proxy-httpd" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.151402 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.164639 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.164922 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.165104 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.181554 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.224115 4847 scope.go:117] "RemoveContainer" containerID="97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.305083 4847 scope.go:117] "RemoveContainer" containerID="e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.315504 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5\": container with ID starting with e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5 not found: ID does not exist" containerID="e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.315552 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5"} err="failed to get container status \"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5\": rpc error: code = NotFound desc = could not find container \"e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5\": container with ID starting with e81b6b4603e3f93c0cf06b5638699f516038225bfc76b85d71baf9cdc24b24e5 not found: ID does not exist" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.315593 4847 scope.go:117] "RemoveContainer" containerID="f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.326881 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9\": container with ID starting with f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9 not found: ID does not exist" containerID="f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.326936 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9"} err="failed to get container status \"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9\": rpc error: code = NotFound desc = could not find container \"f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9\": container with ID starting with f3dd06ad955e855c8a20f436621d008a3181acf27354dd867c76076bf5492ff9 not found: ID does not exist" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327001 4847 scope.go:117] "RemoveContainer" containerID="9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327016 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327098 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327129 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327190 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327222 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327267 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgc6r\" (UniqueName: \"kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327321 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.327347 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.335871 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec\": container with ID starting with 9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec not found: ID does not exist" containerID="9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.335920 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec"} err="failed to get container status \"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec\": rpc error: code = NotFound desc = could not find container \"9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec\": container with ID starting with 9f9cf42f9390c75d6595ae13baded850a7a2b84e423d25add21e3fa77af37bec not found: ID does not exist" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.335958 4847 scope.go:117] "RemoveContainer" containerID="97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1" Dec 10 14:45:20 crc kubenswrapper[4847]: E1210 14:45:20.339867 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1\": container with ID starting with 97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1 not found: ID does not exist" containerID="97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.339914 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1"} err="failed to get container status \"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1\": rpc error: code = NotFound desc = could not find container \"97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1\": container with ID starting with 97cd9e4460bd673839e938b3937427fb18458f588e963a3d9b73cb0c6eae15c1 not found: ID does not exist" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.428831 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429016 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429084 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429115 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429154 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429186 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429227 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgc6r\" (UniqueName: \"kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.429287 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.430111 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.430163 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.434512 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.435420 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.441931 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.441988 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.451047 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgc6r\" (UniqueName: \"kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.451337 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.483601 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.775833 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c74ba68c-4c24-488f-a863-e35cece9f1d9" path="/var/lib/kubelet/pods/c74ba68c-4c24-488f-a863-e35cece9f1d9/volumes" Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.955460 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:20 crc kubenswrapper[4847]: I1210 14:45:20.959895 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 10 14:45:20 crc kubenswrapper[4847]: W1210 14:45:20.963511 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab5fd0ae_b1ec_4fe3_9771_90961891e21e.slice/crio-00374ba2c5a5ac8281ec6a477f50b1ee55b8facb86a61857594f1287e3a586ec WatchSource:0}: Error finding container 00374ba2c5a5ac8281ec6a477f50b1ee55b8facb86a61857594f1287e3a586ec: Status 404 returned error can't find the container with id 00374ba2c5a5ac8281ec6a477f50b1ee55b8facb86a61857594f1287e3a586ec Dec 10 14:45:21 crc kubenswrapper[4847]: I1210 14:45:21.702324 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerStarted","Data":"00374ba2c5a5ac8281ec6a477f50b1ee55b8facb86a61857594f1287e3a586ec"} Dec 10 14:45:22 crc kubenswrapper[4847]: I1210 14:45:22.714229 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerStarted","Data":"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8"} Dec 10 14:45:22 crc kubenswrapper[4847]: I1210 14:45:22.714852 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerStarted","Data":"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226"} Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.054440 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.054884 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.102976 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.153633 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.733509 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerStarted","Data":"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46"} Dec 10 14:45:24 crc kubenswrapper[4847]: I1210 14:45:24.772461 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 14:45:25 crc kubenswrapper[4847]: I1210 14:45:25.137942 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:25 crc kubenswrapper[4847]: I1210 14:45:25.138031 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 14:45:26 crc kubenswrapper[4847]: I1210 14:45:26.755187 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerStarted","Data":"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a"} Dec 10 14:45:26 crc kubenswrapper[4847]: I1210 14:45:26.755967 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:45:26 crc kubenswrapper[4847]: I1210 14:45:26.789858 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.730223537 podStartE2EDuration="6.789832274s" podCreationTimestamp="2025-12-10 14:45:20 +0000 UTC" firstStartedPulling="2025-12-10 14:45:20.966069362 +0000 UTC m=+1270.535286992" lastFinishedPulling="2025-12-10 14:45:26.025678099 +0000 UTC m=+1275.594895729" observedRunningTime="2025-12-10 14:45:26.783511109 +0000 UTC m=+1276.352728739" watchObservedRunningTime="2025-12-10 14:45:26.789832274 +0000 UTC m=+1276.359049904" Dec 10 14:45:27 crc kubenswrapper[4847]: I1210 14:45:27.073308 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 14:45:28 crc kubenswrapper[4847]: I1210 14:45:28.990703 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 14:45:28 crc kubenswrapper[4847]: I1210 14:45:28.995411 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 14:45:28 crc kubenswrapper[4847]: I1210 14:45:28.996131 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 14:45:29 crc kubenswrapper[4847]: I1210 14:45:29.786580 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.713083 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.797562 4847 generic.go:334] "Generic (PLEG): container finished" podID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" containerID="9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0" exitCode=137 Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.797603 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.797611 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03","Type":"ContainerDied","Data":"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0"} Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.797667 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03","Type":"ContainerDied","Data":"b45bbfd19b8adcf4cee809ed5f3ccc230e7647a213ae49c394e88d94c5270845"} Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.797686 4847 scope.go:117] "RemoveContainer" containerID="9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.829546 4847 scope.go:117] "RemoveContainer" containerID="9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0" Dec 10 14:45:30 crc kubenswrapper[4847]: E1210 14:45:30.831007 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0\": container with ID starting with 9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0 not found: ID does not exist" containerID="9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.831047 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0"} err="failed to get container status \"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0\": rpc error: code = NotFound desc = could not find container \"9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0\": container with ID starting with 9e6eb44949766b7c580341c8a4401fb0acea906409fbcbb317f5bc082bf055b0 not found: ID does not exist" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.831668 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5b87\" (UniqueName: \"kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87\") pod \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.831861 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data\") pod \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.831911 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle\") pod \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\" (UID: \"95f88c35-2ac4-47d6-b4dc-21dfd14d8e03\") " Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.848384 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87" (OuterVolumeSpecName: "kube-api-access-q5b87") pod "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" (UID: "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03"). InnerVolumeSpecName "kube-api-access-q5b87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.865121 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" (UID: "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.880067 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data" (OuterVolumeSpecName: "config-data") pod "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" (UID: "95f88c35-2ac4-47d6-b4dc-21dfd14d8e03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.934607 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5b87\" (UniqueName: \"kubernetes.io/projected/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-kube-api-access-q5b87\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.934643 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:30 crc kubenswrapper[4847]: I1210 14:45:30.934653 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.010694 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.010769 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.128781 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.138903 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.155183 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:45:31 crc kubenswrapper[4847]: E1210 14:45:31.155626 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.155649 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.155952 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.159077 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.161201 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.161467 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.161815 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.171192 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.240219 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.240614 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcfk7\" (UniqueName: \"kubernetes.io/projected/d1d12a43-2aba-411a-ad3d-628d6f468c95-kube-api-access-zcfk7\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.240839 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.240989 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.241159 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.344090 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.344525 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.344655 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.344834 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.345002 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcfk7\" (UniqueName: \"kubernetes.io/projected/d1d12a43-2aba-411a-ad3d-628d6f468c95-kube-api-access-zcfk7\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.352432 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.352700 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.353692 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.353756 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d12a43-2aba-411a-ad3d-628d6f468c95-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.366288 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcfk7\" (UniqueName: \"kubernetes.io/projected/d1d12a43-2aba-411a-ad3d-628d6f468c95-kube-api-access-zcfk7\") pod \"nova-cell1-novncproxy-0\" (UID: \"d1d12a43-2aba-411a-ad3d-628d6f468c95\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.481471 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:31 crc kubenswrapper[4847]: I1210 14:45:31.958300 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 14:45:31 crc kubenswrapper[4847]: W1210 14:45:31.968402 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1d12a43_2aba_411a_ad3d_628d6f468c95.slice/crio-949c7abe15ba1cdc3dc38607650c32e77bd7956b30ea44d8e7aa818519c718c2 WatchSource:0}: Error finding container 949c7abe15ba1cdc3dc38607650c32e77bd7956b30ea44d8e7aa818519c718c2: Status 404 returned error can't find the container with id 949c7abe15ba1cdc3dc38607650c32e77bd7956b30ea44d8e7aa818519c718c2 Dec 10 14:45:32 crc kubenswrapper[4847]: I1210 14:45:32.773167 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f88c35-2ac4-47d6-b4dc-21dfd14d8e03" path="/var/lib/kubelet/pods/95f88c35-2ac4-47d6-b4dc-21dfd14d8e03/volumes" Dec 10 14:45:32 crc kubenswrapper[4847]: I1210 14:45:32.829928 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d1d12a43-2aba-411a-ad3d-628d6f468c95","Type":"ContainerStarted","Data":"78a09130275c8c39bbb3b94a213996ac3abf66c95122f9a327e4eabc35540485"} Dec 10 14:45:32 crc kubenswrapper[4847]: I1210 14:45:32.829974 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d1d12a43-2aba-411a-ad3d-628d6f468c95","Type":"ContainerStarted","Data":"949c7abe15ba1cdc3dc38607650c32e77bd7956b30ea44d8e7aa818519c718c2"} Dec 10 14:45:32 crc kubenswrapper[4847]: I1210 14:45:32.847549 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.8475257059999999 podStartE2EDuration="1.847525706s" podCreationTimestamp="2025-12-10 14:45:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:32.845145464 +0000 UTC m=+1282.414363104" watchObservedRunningTime="2025-12-10 14:45:32.847525706 +0000 UTC m=+1282.416743336" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.060200 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.060901 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.065645 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.077012 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.850085 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 14:45:34 crc kubenswrapper[4847]: I1210 14:45:34.854233 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.052264 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.054138 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.097853 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.123636 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.123789 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.123830 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.123893 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfsg4\" (UniqueName: \"kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.123925 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.124289 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226202 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226270 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226330 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfsg4\" (UniqueName: \"kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226371 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226440 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.226476 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.227359 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.227423 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.227437 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.227585 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.228281 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.248946 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfsg4\" (UniqueName: \"kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4\") pod \"dnsmasq-dns-cd5cbd7b9-dd2sj\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:35 crc kubenswrapper[4847]: I1210 14:45:35.397696 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:36 crc kubenswrapper[4847]: I1210 14:45:36.011658 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:45:36 crc kubenswrapper[4847]: I1210 14:45:36.482675 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:36 crc kubenswrapper[4847]: I1210 14:45:36.868504 4847 generic.go:334] "Generic (PLEG): container finished" podID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerID="066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9" exitCode=0 Dec 10 14:45:36 crc kubenswrapper[4847]: I1210 14:45:36.868584 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" event={"ID":"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb","Type":"ContainerDied","Data":"066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9"} Dec 10 14:45:36 crc kubenswrapper[4847]: I1210 14:45:36.868647 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" event={"ID":"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb","Type":"ContainerStarted","Data":"adf77893fad6ccba89507973cb11cd6be97170c067f79acb132aafc4f2df7cf8"} Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.456971 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.457802 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-central-agent" containerID="cri-o://b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.457917 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="proxy-httpd" containerID="cri-o://2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.457957 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="sg-core" containerID="cri-o://57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.458004 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-notification-agent" containerID="cri-o://11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.484484 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.619042 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.887253 4847 generic.go:334] "Generic (PLEG): container finished" podID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerID="2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a" exitCode=0 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.887564 4847 generic.go:334] "Generic (PLEG): container finished" podID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerID="57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46" exitCode=2 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.887305 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerDied","Data":"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a"} Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.887911 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerDied","Data":"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46"} Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.890490 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-log" containerID="cri-o://788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.891197 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" event={"ID":"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb","Type":"ContainerStarted","Data":"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36"} Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.891279 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-api" containerID="cri-o://d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4" gracePeriod=30 Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.891518 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:37 crc kubenswrapper[4847]: I1210 14:45:37.917730 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" podStartSLOduration=2.917689227 podStartE2EDuration="2.917689227s" podCreationTimestamp="2025-12-10 14:45:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:37.914560375 +0000 UTC m=+1287.483778015" watchObservedRunningTime="2025-12-10 14:45:37.917689227 +0000 UTC m=+1287.486906857" Dec 10 14:45:38 crc kubenswrapper[4847]: I1210 14:45:38.913798 4847 generic.go:334] "Generic (PLEG): container finished" podID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerID="788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1" exitCode=143 Dec 10 14:45:38 crc kubenswrapper[4847]: I1210 14:45:38.913934 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerDied","Data":"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1"} Dec 10 14:45:38 crc kubenswrapper[4847]: I1210 14:45:38.923384 4847 generic.go:334] "Generic (PLEG): container finished" podID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerID="b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226" exitCode=0 Dec 10 14:45:38 crc kubenswrapper[4847]: I1210 14:45:38.923502 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerDied","Data":"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226"} Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.385777 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.517300 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgc6r\" (UniqueName: \"kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518042 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518087 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518125 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518182 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518272 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518343 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518427 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml\") pod \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\" (UID: \"ab5fd0ae-b1ec-4fe3-9771-90961891e21e\") " Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518624 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.518797 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.519111 4847 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.519137 4847 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.539189 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts" (OuterVolumeSpecName: "scripts") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.556925 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r" (OuterVolumeSpecName: "kube-api-access-vgc6r") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "kube-api-access-vgc6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.603121 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.626204 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.626241 4847 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.626250 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgc6r\" (UniqueName: \"kubernetes.io/projected/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-kube-api-access-vgc6r\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.695905 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.718606 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data" (OuterVolumeSpecName: "config-data") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.728120 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.728167 4847 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.736865 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab5fd0ae-b1ec-4fe3-9771-90961891e21e" (UID: "ab5fd0ae-b1ec-4fe3-9771-90961891e21e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.830279 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5fd0ae-b1ec-4fe3-9771-90961891e21e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.938689 4847 generic.go:334] "Generic (PLEG): container finished" podID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerID="11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8" exitCode=0 Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.938748 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerDied","Data":"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8"} Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.938780 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.938791 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ab5fd0ae-b1ec-4fe3-9771-90961891e21e","Type":"ContainerDied","Data":"00374ba2c5a5ac8281ec6a477f50b1ee55b8facb86a61857594f1287e3a586ec"} Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.938812 4847 scope.go:117] "RemoveContainer" containerID="2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.962994 4847 scope.go:117] "RemoveContainer" containerID="57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46" Dec 10 14:45:39 crc kubenswrapper[4847]: I1210 14:45:39.992027 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.007479 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.017875 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.018437 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-notification-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.018500 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-notification-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.018561 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="sg-core" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.018610 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="sg-core" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.018671 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-central-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.018742 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-central-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.018815 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="proxy-httpd" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.018864 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="proxy-httpd" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.019075 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-central-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.019155 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="sg-core" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.019206 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="proxy-httpd" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.019261 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" containerName="ceilometer-notification-agent" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.024941 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.028871 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.029183 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.029437 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.044875 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.051015 4847 scope.go:117] "RemoveContainer" containerID="11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.081641 4847 scope.go:117] "RemoveContainer" containerID="b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.109620 4847 scope.go:117] "RemoveContainer" containerID="2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.110545 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a\": container with ID starting with 2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a not found: ID does not exist" containerID="2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.110584 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a"} err="failed to get container status \"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a\": rpc error: code = NotFound desc = could not find container \"2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a\": container with ID starting with 2c26d98d345abfb6c034023e31eadb4a7f51dfb21356126a816bdbe4e631e95a not found: ID does not exist" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.110609 4847 scope.go:117] "RemoveContainer" containerID="57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.111066 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46\": container with ID starting with 57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46 not found: ID does not exist" containerID="57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.111107 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46"} err="failed to get container status \"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46\": rpc error: code = NotFound desc = could not find container \"57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46\": container with ID starting with 57a61fd1fa1a9ad8876eb223a0f7ac0c3bc61caa1b6df8f7df227901403d3b46 not found: ID does not exist" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.111135 4847 scope.go:117] "RemoveContainer" containerID="11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.111493 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8\": container with ID starting with 11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8 not found: ID does not exist" containerID="11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.111519 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8"} err="failed to get container status \"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8\": rpc error: code = NotFound desc = could not find container \"11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8\": container with ID starting with 11be5d63a961bf19083d3b02c65b6c160185387350b69d675eb3326e18b3e6d8 not found: ID does not exist" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.111537 4847 scope.go:117] "RemoveContainer" containerID="b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226" Dec 10 14:45:40 crc kubenswrapper[4847]: E1210 14:45:40.111854 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226\": container with ID starting with b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226 not found: ID does not exist" containerID="b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.111889 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226"} err="failed to get container status \"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226\": rpc error: code = NotFound desc = could not find container \"b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226\": container with ID starting with b426983c69580c5bab195e803565528cdec40cade5ab2d593c74ac92ff855226 not found: ID does not exist" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.136931 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137016 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-config-data\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-log-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137102 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-scripts\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137134 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137165 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137207 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-run-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.137381 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2rb2\" (UniqueName: \"kubernetes.io/projected/3cd44098-3d49-408d-8478-fceaa9ac7e57-kube-api-access-x2rb2\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239665 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239751 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-config-data\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239775 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-log-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239798 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-scripts\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239820 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239841 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239869 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-run-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.239908 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2rb2\" (UniqueName: \"kubernetes.io/projected/3cd44098-3d49-408d-8478-fceaa9ac7e57-kube-api-access-x2rb2\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.242296 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-run-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.242322 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3cd44098-3d49-408d-8478-fceaa9ac7e57-log-httpd\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.246639 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-scripts\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.248195 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.251218 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.256018 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-config-data\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.258101 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd44098-3d49-408d-8478-fceaa9ac7e57-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.258865 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2rb2\" (UniqueName: \"kubernetes.io/projected/3cd44098-3d49-408d-8478-fceaa9ac7e57-kube-api-access-x2rb2\") pod \"ceilometer-0\" (UID: \"3cd44098-3d49-408d-8478-fceaa9ac7e57\") " pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.369161 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.771012 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab5fd0ae-b1ec-4fe3-9771-90961891e21e" path="/var/lib/kubelet/pods/ab5fd0ae-b1ec-4fe3-9771-90961891e21e/volumes" Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.830002 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 14:45:40 crc kubenswrapper[4847]: W1210 14:45:40.831721 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cd44098_3d49_408d_8478_fceaa9ac7e57.slice/crio-0a1f17ac132e48538bed21b3470a55148bbba30af63574f432627e959219d5b0 WatchSource:0}: Error finding container 0a1f17ac132e48538bed21b3470a55148bbba30af63574f432627e959219d5b0: Status 404 returned error can't find the container with id 0a1f17ac132e48538bed21b3470a55148bbba30af63574f432627e959219d5b0 Dec 10 14:45:40 crc kubenswrapper[4847]: I1210 14:45:40.957019 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cd44098-3d49-408d-8478-fceaa9ac7e57","Type":"ContainerStarted","Data":"0a1f17ac132e48538bed21b3470a55148bbba30af63574f432627e959219d5b0"} Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.481853 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.503431 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.513620 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.667942 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle\") pod \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.668083 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs\") pod \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.668151 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data\") pod \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.668312 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whjzr\" (UniqueName: \"kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr\") pod \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\" (UID: \"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9\") " Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.669667 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs" (OuterVolumeSpecName: "logs") pod "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" (UID: "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.675784 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr" (OuterVolumeSpecName: "kube-api-access-whjzr") pod "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" (UID: "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9"). InnerVolumeSpecName "kube-api-access-whjzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.700875 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data" (OuterVolumeSpecName: "config-data") pod "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" (UID: "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.705608 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" (UID: "3929c15a-fe24-4c13-bf2e-e9264bbcdfe9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.770756 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.771839 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whjzr\" (UniqueName: \"kubernetes.io/projected/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-kube-api-access-whjzr\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.772527 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.772540 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.967632 4847 generic.go:334] "Generic (PLEG): container finished" podID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerID="d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4" exitCode=0 Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.967740 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerDied","Data":"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4"} Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.967774 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3929c15a-fe24-4c13-bf2e-e9264bbcdfe9","Type":"ContainerDied","Data":"cf05691bf39ddb00d00bc1bb19eba0c10edad39ca88420912e8acffe5fa9368c"} Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.967798 4847 scope.go:117] "RemoveContainer" containerID="d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.967822 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.973378 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cd44098-3d49-408d-8478-fceaa9ac7e57","Type":"ContainerStarted","Data":"d356b496191479326375e2c16b15e241dbabf519048b9dba626d58f1daec56f1"} Dec 10 14:45:41 crc kubenswrapper[4847]: I1210 14:45:41.987957 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.047539 4847 scope.go:117] "RemoveContainer" containerID="788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.051649 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.065700 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.090341 4847 scope.go:117] "RemoveContainer" containerID="d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4" Dec 10 14:45:42 crc kubenswrapper[4847]: E1210 14:45:42.091135 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4\": container with ID starting with d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4 not found: ID does not exist" containerID="d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.091246 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4"} err="failed to get container status \"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4\": rpc error: code = NotFound desc = could not find container \"d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4\": container with ID starting with d1cef7951c5b7eb1535c19cc2a347b1f573c5448b91fa285069527b50bfc55e4 not found: ID does not exist" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.091357 4847 scope.go:117] "RemoveContainer" containerID="788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1" Dec 10 14:45:42 crc kubenswrapper[4847]: E1210 14:45:42.092953 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1\": container with ID starting with 788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1 not found: ID does not exist" containerID="788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.093006 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1"} err="failed to get container status \"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1\": rpc error: code = NotFound desc = could not find container \"788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1\": container with ID starting with 788364ec9e23ddeeaa2f92709ab470a522d7fc0a5a68970a7e32a0a9bc4294f1 not found: ID does not exist" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.121555 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:42 crc kubenswrapper[4847]: E1210 14:45:42.122071 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-log" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.122093 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-log" Dec 10 14:45:42 crc kubenswrapper[4847]: E1210 14:45:42.122112 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-api" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.122121 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-api" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.122341 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-api" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.122369 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" containerName="nova-api-log" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.128106 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.138258 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.142667 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.143040 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.143078 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.182250 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.182307 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.182595 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.182670 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhlsf\" (UniqueName: \"kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.182847 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.183222 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.266932 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pwcls"] Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.268662 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.275266 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.275641 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.284317 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pwcls"] Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285551 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285606 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhlsf\" (UniqueName: \"kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285664 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285938 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.285985 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.286457 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.295561 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.297049 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.298275 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.299246 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.315443 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhlsf\" (UniqueName: \"kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf\") pod \"nova-api-0\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.388345 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.388555 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.388744 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.388788 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9whp\" (UniqueName: \"kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.455071 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.490417 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.490548 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.490581 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9whp\" (UniqueName: \"kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.490601 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.494797 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.495834 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.496065 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.511859 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9whp\" (UniqueName: \"kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp\") pod \"nova-cell1-cell-mapping-pwcls\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.590605 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.786755 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3929c15a-fe24-4c13-bf2e-e9264bbcdfe9" path="/var/lib/kubelet/pods/3929c15a-fe24-4c13-bf2e-e9264bbcdfe9/volumes" Dec 10 14:45:42 crc kubenswrapper[4847]: I1210 14:45:42.982898 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cd44098-3d49-408d-8478-fceaa9ac7e57","Type":"ContainerStarted","Data":"6c4b6afa101d75eb893f156f63ac555f4a8c14a51ac857c085cc0ea116fb6d40"} Dec 10 14:45:43 crc kubenswrapper[4847]: I1210 14:45:43.010838 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:43 crc kubenswrapper[4847]: W1210 14:45:43.018616 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde1f408b_42ab_40f0_bdee_2aa68e73f5b1.slice/crio-505238efb55a228e0eb84aa850ed775b2aa32f6cce3fea4711a4efac70e8b894 WatchSource:0}: Error finding container 505238efb55a228e0eb84aa850ed775b2aa32f6cce3fea4711a4efac70e8b894: Status 404 returned error can't find the container with id 505238efb55a228e0eb84aa850ed775b2aa32f6cce3fea4711a4efac70e8b894 Dec 10 14:45:43 crc kubenswrapper[4847]: I1210 14:45:43.106747 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pwcls"] Dec 10 14:45:43 crc kubenswrapper[4847]: W1210 14:45:43.113123 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb71e2f0_928f_4d58_adf9_c5127aa279c5.slice/crio-66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c WatchSource:0}: Error finding container 66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c: Status 404 returned error can't find the container with id 66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.002277 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerStarted","Data":"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.002657 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerStarted","Data":"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.002677 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerStarted","Data":"505238efb55a228e0eb84aa850ed775b2aa32f6cce3fea4711a4efac70e8b894"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.008300 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pwcls" event={"ID":"bb71e2f0-928f-4d58-adf9-c5127aa279c5","Type":"ContainerStarted","Data":"098cffea187b81968549c2a1ae43d92cb3fc93d790fecaa25464ab888776a178"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.008348 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pwcls" event={"ID":"bb71e2f0-928f-4d58-adf9-c5127aa279c5","Type":"ContainerStarted","Data":"66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.012527 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cd44098-3d49-408d-8478-fceaa9ac7e57","Type":"ContainerStarted","Data":"978013b36a347333bdd1cafae3f754aca0fe716b2ede0ba5c4a9f3155b2700e1"} Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.036610 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.036585403 podStartE2EDuration="2.036585403s" podCreationTimestamp="2025-12-10 14:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:44.023093901 +0000 UTC m=+1293.592311551" watchObservedRunningTime="2025-12-10 14:45:44.036585403 +0000 UTC m=+1293.605803033" Dec 10 14:45:44 crc kubenswrapper[4847]: I1210 14:45:44.051957 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pwcls" podStartSLOduration=2.051935852 podStartE2EDuration="2.051935852s" podCreationTimestamp="2025-12-10 14:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:44.048244376 +0000 UTC m=+1293.617462026" watchObservedRunningTime="2025-12-10 14:45:44.051935852 +0000 UTC m=+1293.621153472" Dec 10 14:45:45 crc kubenswrapper[4847]: I1210 14:45:45.400863 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:45:45 crc kubenswrapper[4847]: I1210 14:45:45.475480 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:45:45 crc kubenswrapper[4847]: I1210 14:45:45.475935 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="dnsmasq-dns" containerID="cri-o://d87007198c6c559570e1e0a3e0bb6b1acc336f37eb62d921a078a4c9a387f7df" gracePeriod=10 Dec 10 14:45:45 crc kubenswrapper[4847]: I1210 14:45:45.723615 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.186:5353: connect: connection refused" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.040213 4847 generic.go:334] "Generic (PLEG): container finished" podID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerID="d87007198c6c559570e1e0a3e0bb6b1acc336f37eb62d921a078a4c9a387f7df" exitCode=0 Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.040275 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" event={"ID":"037939fa-5737-4f3e-8a25-b7532272d5f1","Type":"ContainerDied","Data":"d87007198c6c559570e1e0a3e0bb6b1acc336f37eb62d921a078a4c9a387f7df"} Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.046978 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3cd44098-3d49-408d-8478-fceaa9ac7e57","Type":"ContainerStarted","Data":"1eb7ed67478c1d101a79ea26758aff77c575cd82f7bb3d8cbad95c1e874ef817"} Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.048177 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.066561 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8158234049999997 podStartE2EDuration="7.066544062s" podCreationTimestamp="2025-12-10 14:45:39 +0000 UTC" firstStartedPulling="2025-12-10 14:45:40.834399734 +0000 UTC m=+1290.403617354" lastFinishedPulling="2025-12-10 14:45:45.085120381 +0000 UTC m=+1294.654338011" observedRunningTime="2025-12-10 14:45:46.065248679 +0000 UTC m=+1295.634466319" watchObservedRunningTime="2025-12-10 14:45:46.066544062 +0000 UTC m=+1295.635761692" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.505599 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583324 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583402 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583486 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583516 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583605 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.583627 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj8bk\" (UniqueName: \"kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk\") pod \"037939fa-5737-4f3e-8a25-b7532272d5f1\" (UID: \"037939fa-5737-4f3e-8a25-b7532272d5f1\") " Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.598050 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk" (OuterVolumeSpecName: "kube-api-access-wj8bk") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "kube-api-access-wj8bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.643595 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.656449 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.666669 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.674220 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.685789 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.685825 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.685837 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.685849 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.685865 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj8bk\" (UniqueName: \"kubernetes.io/projected/037939fa-5737-4f3e-8a25-b7532272d5f1-kube-api-access-wj8bk\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.700523 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config" (OuterVolumeSpecName: "config") pod "037939fa-5737-4f3e-8a25-b7532272d5f1" (UID: "037939fa-5737-4f3e-8a25-b7532272d5f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:45:46 crc kubenswrapper[4847]: I1210 14:45:46.787282 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037939fa-5737-4f3e-8a25-b7532272d5f1-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.064486 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.065001 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-wwn4z" event={"ID":"037939fa-5737-4f3e-8a25-b7532272d5f1","Type":"ContainerDied","Data":"af64a8b760f9323eb48d24d4c05b44d596840dc5ac4dff76cfd566fc3f59f99c"} Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.065032 4847 scope.go:117] "RemoveContainer" containerID="d87007198c6c559570e1e0a3e0bb6b1acc336f37eb62d921a078a4c9a387f7df" Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.086676 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.088763 4847 scope.go:117] "RemoveContainer" containerID="2f08f2148881fdb195190e1ef241a8177fd83a607a62c6a96844e52e4abaceca" Dec 10 14:45:47 crc kubenswrapper[4847]: I1210 14:45:47.097288 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-wwn4z"] Dec 10 14:45:48 crc kubenswrapper[4847]: I1210 14:45:48.783636 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" path="/var/lib/kubelet/pods/037939fa-5737-4f3e-8a25-b7532272d5f1/volumes" Dec 10 14:45:49 crc kubenswrapper[4847]: I1210 14:45:49.083193 4847 generic.go:334] "Generic (PLEG): container finished" podID="bb71e2f0-928f-4d58-adf9-c5127aa279c5" containerID="098cffea187b81968549c2a1ae43d92cb3fc93d790fecaa25464ab888776a178" exitCode=0 Dec 10 14:45:49 crc kubenswrapper[4847]: I1210 14:45:49.083246 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pwcls" event={"ID":"bb71e2f0-928f-4d58-adf9-c5127aa279c5","Type":"ContainerDied","Data":"098cffea187b81968549c2a1ae43d92cb3fc93d790fecaa25464ab888776a178"} Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.469603 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.561353 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data\") pod \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.561563 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts\") pod \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.561643 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9whp\" (UniqueName: \"kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp\") pod \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.561680 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle\") pod \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\" (UID: \"bb71e2f0-928f-4d58-adf9-c5127aa279c5\") " Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.567660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp" (OuterVolumeSpecName: "kube-api-access-t9whp") pod "bb71e2f0-928f-4d58-adf9-c5127aa279c5" (UID: "bb71e2f0-928f-4d58-adf9-c5127aa279c5"). InnerVolumeSpecName "kube-api-access-t9whp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.567919 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts" (OuterVolumeSpecName: "scripts") pod "bb71e2f0-928f-4d58-adf9-c5127aa279c5" (UID: "bb71e2f0-928f-4d58-adf9-c5127aa279c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.591204 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb71e2f0-928f-4d58-adf9-c5127aa279c5" (UID: "bb71e2f0-928f-4d58-adf9-c5127aa279c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.600190 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data" (OuterVolumeSpecName: "config-data") pod "bb71e2f0-928f-4d58-adf9-c5127aa279c5" (UID: "bb71e2f0-928f-4d58-adf9-c5127aa279c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.663901 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.663949 4847 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.663962 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9whp\" (UniqueName: \"kubernetes.io/projected/bb71e2f0-928f-4d58-adf9-c5127aa279c5-kube-api-access-t9whp\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:50 crc kubenswrapper[4847]: I1210 14:45:50.663978 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb71e2f0-928f-4d58-adf9-c5127aa279c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.108129 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pwcls" event={"ID":"bb71e2f0-928f-4d58-adf9-c5127aa279c5","Type":"ContainerDied","Data":"66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c"} Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.108476 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a96edb2803ba1ba44cb4de0ecf86e2180e60d2db602db6b09b759f9708885c" Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.108234 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pwcls" Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.283472 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.283804 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-log" containerID="cri-o://2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" gracePeriod=30 Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.283834 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-api" containerID="cri-o://1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" gracePeriod=30 Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.313893 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.314143 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0d2d1859-c765-4069-9ad6-8c919c5d2783" containerName="nova-scheduler-scheduler" containerID="cri-o://f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7" gracePeriod=30 Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.325077 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.325346 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" containerID="cri-o://d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231" gracePeriod=30 Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.325517 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" containerID="cri-o://204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733" gracePeriod=30 Dec 10 14:45:51 crc kubenswrapper[4847]: I1210 14:45:51.958867 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094216 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhlsf\" (UniqueName: \"kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094294 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094453 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094476 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094593 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.094645 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs\") pod \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\" (UID: \"de1f408b-42ab-40f0-bdee-2aa68e73f5b1\") " Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.095368 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs" (OuterVolumeSpecName: "logs") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.100610 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf" (OuterVolumeSpecName: "kube-api-access-zhlsf") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "kube-api-access-zhlsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.121394 4847 generic.go:334] "Generic (PLEG): container finished" podID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerID="d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231" exitCode=143 Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.121474 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerDied","Data":"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231"} Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125123 4847 generic.go:334] "Generic (PLEG): container finished" podID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerID="1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" exitCode=0 Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125158 4847 generic.go:334] "Generic (PLEG): container finished" podID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerID="2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" exitCode=143 Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125185 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerDied","Data":"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e"} Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125221 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerDied","Data":"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d"} Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125235 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de1f408b-42ab-40f0-bdee-2aa68e73f5b1","Type":"ContainerDied","Data":"505238efb55a228e0eb84aa850ed775b2aa32f6cce3fea4711a4efac70e8b894"} Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125256 4847 scope.go:117] "RemoveContainer" containerID="1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.125439 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.128100 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data" (OuterVolumeSpecName: "config-data") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.146382 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.153155 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.157613 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "de1f408b-42ab-40f0-bdee-2aa68e73f5b1" (UID: "de1f408b-42ab-40f0-bdee-2aa68e73f5b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.157685 4847 scope.go:117] "RemoveContainer" containerID="2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.183307 4847 scope.go:117] "RemoveContainer" containerID="1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.185349 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e\": container with ID starting with 1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e not found: ID does not exist" containerID="1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.185445 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e"} err="failed to get container status \"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e\": rpc error: code = NotFound desc = could not find container \"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e\": container with ID starting with 1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e not found: ID does not exist" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.185553 4847 scope.go:117] "RemoveContainer" containerID="2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.186018 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d\": container with ID starting with 2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d not found: ID does not exist" containerID="2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.186045 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d"} err="failed to get container status \"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d\": rpc error: code = NotFound desc = could not find container \"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d\": container with ID starting with 2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d not found: ID does not exist" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.186061 4847 scope.go:117] "RemoveContainer" containerID="1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.186418 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e"} err="failed to get container status \"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e\": rpc error: code = NotFound desc = could not find container \"1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e\": container with ID starting with 1a4977ec8587bc9702cdd13e1a2bec14b65da29f2d310fab8a03cd989ba8459e not found: ID does not exist" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.186460 4847 scope.go:117] "RemoveContainer" containerID="2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.186699 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d"} err="failed to get container status \"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d\": rpc error: code = NotFound desc = could not find container \"2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d\": container with ID starting with 2b76904e515252446687b1871b2d22398ac936d6289685f9693b1915c5e8a81d not found: ID does not exist" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197196 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197238 4847 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197250 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197286 4847 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197295 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhlsf\" (UniqueName: \"kubernetes.io/projected/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-kube-api-access-zhlsf\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.197303 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de1f408b-42ab-40f0-bdee-2aa68e73f5b1-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.468630 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.480633 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496174 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.496656 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb71e2f0-928f-4d58-adf9-c5127aa279c5" containerName="nova-manage" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496684 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb71e2f0-928f-4d58-adf9-c5127aa279c5" containerName="nova-manage" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.496702 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="dnsmasq-dns" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496730 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="dnsmasq-dns" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.496742 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-log" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496751 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-log" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.496769 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-api" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496777 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-api" Dec 10 14:45:52 crc kubenswrapper[4847]: E1210 14:45:52.496810 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="init" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.496820 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="init" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.497108 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-log" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.497136 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" containerName="nova-api-api" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.497169 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="037939fa-5737-4f3e-8a25-b7532272d5f1" containerName="dnsmasq-dns" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.497182 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb71e2f0-928f-4d58-adf9-c5127aa279c5" containerName="nova-manage" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.498286 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.501314 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.501545 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.501678 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.516466 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.604512 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-logs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.604585 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.604689 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.604825 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.604938 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-config-data\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.605065 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmvlq\" (UniqueName: \"kubernetes.io/projected/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-kube-api-access-cmvlq\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.706891 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707003 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707057 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707103 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-config-data\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707160 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmvlq\" (UniqueName: \"kubernetes.io/projected/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-kube-api-access-cmvlq\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707184 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-logs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.707578 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-logs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.711096 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.711388 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.711242 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-config-data\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.716363 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.722485 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmvlq\" (UniqueName: \"kubernetes.io/projected/0205d3f1-fa73-4291-a22b-1bdd4aaa23ec-kube-api-access-cmvlq\") pod \"nova-api-0\" (UID: \"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec\") " pod="openstack/nova-api-0" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.774118 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1f408b-42ab-40f0-bdee-2aa68e73f5b1" path="/var/lib/kubelet/pods/de1f408b-42ab-40f0-bdee-2aa68e73f5b1/volumes" Dec 10 14:45:52 crc kubenswrapper[4847]: I1210 14:45:52.827102 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.298042 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 14:45:53 crc kubenswrapper[4847]: W1210 14:45:53.301306 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0205d3f1_fa73_4291_a22b_1bdd4aaa23ec.slice/crio-b480953cdd3d6ac82a7e8ce23bbb391f3c4019863a6512725cdf234a3639ec97 WatchSource:0}: Error finding container b480953cdd3d6ac82a7e8ce23bbb391f3c4019863a6512725cdf234a3639ec97: Status 404 returned error can't find the container with id b480953cdd3d6ac82a7e8ce23bbb391f3c4019863a6512725cdf234a3639ec97 Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.848008 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.935055 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data\") pod \"0d2d1859-c765-4069-9ad6-8c919c5d2783\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.935760 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle\") pod \"0d2d1859-c765-4069-9ad6-8c919c5d2783\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.935957 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4fdf\" (UniqueName: \"kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf\") pod \"0d2d1859-c765-4069-9ad6-8c919c5d2783\" (UID: \"0d2d1859-c765-4069-9ad6-8c919c5d2783\") " Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.941626 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf" (OuterVolumeSpecName: "kube-api-access-w4fdf") pod "0d2d1859-c765-4069-9ad6-8c919c5d2783" (UID: "0d2d1859-c765-4069-9ad6-8c919c5d2783"). InnerVolumeSpecName "kube-api-access-w4fdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.964681 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d2d1859-c765-4069-9ad6-8c919c5d2783" (UID: "0d2d1859-c765-4069-9ad6-8c919c5d2783"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:53 crc kubenswrapper[4847]: I1210 14:45:53.965264 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data" (OuterVolumeSpecName: "config-data") pod "0d2d1859-c765-4069-9ad6-8c919c5d2783" (UID: "0d2d1859-c765-4069-9ad6-8c919c5d2783"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.039199 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.039237 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2d1859-c765-4069-9ad6-8c919c5d2783-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.039250 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4fdf\" (UniqueName: \"kubernetes.io/projected/0d2d1859-c765-4069-9ad6-8c919c5d2783-kube-api-access-w4fdf\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.175036 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec","Type":"ContainerStarted","Data":"63d7f61b3ddd3755c2c1c49b99c52ae96d41d76ce1c45dac07c1c8a3967610db"} Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.175092 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec","Type":"ContainerStarted","Data":"479c89d78a53b61ca0d08a671b401524ce36493c7ca851f47c62035e2758efdc"} Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.175102 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0205d3f1-fa73-4291-a22b-1bdd4aaa23ec","Type":"ContainerStarted","Data":"b480953cdd3d6ac82a7e8ce23bbb391f3c4019863a6512725cdf234a3639ec97"} Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.181136 4847 generic.go:334] "Generic (PLEG): container finished" podID="0d2d1859-c765-4069-9ad6-8c919c5d2783" containerID="f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7" exitCode=0 Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.181179 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d2d1859-c765-4069-9ad6-8c919c5d2783","Type":"ContainerDied","Data":"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7"} Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.181205 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0d2d1859-c765-4069-9ad6-8c919c5d2783","Type":"ContainerDied","Data":"d4fb498287540bad39aa853cb6d7cba4c6d482c43dc451dc665a4f234306b646"} Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.181223 4847 scope.go:117] "RemoveContainer" containerID="f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.181367 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.224231 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.224207457 podStartE2EDuration="2.224207457s" podCreationTimestamp="2025-12-10 14:45:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:54.211533547 +0000 UTC m=+1303.780751187" watchObservedRunningTime="2025-12-10 14:45:54.224207457 +0000 UTC m=+1303.793425087" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.228785 4847 scope.go:117] "RemoveContainer" containerID="f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7" Dec 10 14:45:54 crc kubenswrapper[4847]: E1210 14:45:54.229249 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7\": container with ID starting with f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7 not found: ID does not exist" containerID="f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.229277 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7"} err="failed to get container status \"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7\": rpc error: code = NotFound desc = could not find container \"f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7\": container with ID starting with f0670d6753ba84314fcd50a0e4501305276e46965515822de17535ae215f01c7 not found: ID does not exist" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.246204 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.269213 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.282876 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:54 crc kubenswrapper[4847]: E1210 14:45:54.283268 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2d1859-c765-4069-9ad6-8c919c5d2783" containerName="nova-scheduler-scheduler" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.283288 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2d1859-c765-4069-9ad6-8c919c5d2783" containerName="nova-scheduler-scheduler" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.283466 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2d1859-c765-4069-9ad6-8c919c5d2783" containerName="nova-scheduler-scheduler" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.284127 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.287570 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.292157 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.345043 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qzpq\" (UniqueName: \"kubernetes.io/projected/8b228613-6a71-4c3b-a95a-e0a496c0b346-kube-api-access-6qzpq\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.345399 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.345442 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-config-data\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.448302 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-config-data\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.449349 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qzpq\" (UniqueName: \"kubernetes.io/projected/8b228613-6a71-4c3b-a95a-e0a496c0b346-kube-api-access-6qzpq\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.449401 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.455413 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.455846 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b228613-6a71-4c3b-a95a-e0a496c0b346-config-data\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.475430 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:49498->10.217.0.190:8775: read: connection reset by peer" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.475448 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:49510->10.217.0.190:8775: read: connection reset by peer" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.476606 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qzpq\" (UniqueName: \"kubernetes.io/projected/8b228613-6a71-4c3b-a95a-e0a496c0b346-kube-api-access-6qzpq\") pod \"nova-scheduler-0\" (UID: \"8b228613-6a71-4c3b-a95a-e0a496c0b346\") " pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.603972 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.775886 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d2d1859-c765-4069-9ad6-8c919c5d2783" path="/var/lib/kubelet/pods/0d2d1859-c765-4069-9ad6-8c919c5d2783/volumes" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.866097 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.960429 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs\") pod \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.960607 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle\") pod \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.960702 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data\") pod \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.960808 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6g9q\" (UniqueName: \"kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q\") pod \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.960863 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs\") pod \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\" (UID: \"216dbcc7-4f75-40eb-9c86-bf4ffff1879f\") " Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.961981 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs" (OuterVolumeSpecName: "logs") pod "216dbcc7-4f75-40eb-9c86-bf4ffff1879f" (UID: "216dbcc7-4f75-40eb-9c86-bf4ffff1879f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.966840 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q" (OuterVolumeSpecName: "kube-api-access-x6g9q") pod "216dbcc7-4f75-40eb-9c86-bf4ffff1879f" (UID: "216dbcc7-4f75-40eb-9c86-bf4ffff1879f"). InnerVolumeSpecName "kube-api-access-x6g9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:45:54 crc kubenswrapper[4847]: I1210 14:45:54.992631 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data" (OuterVolumeSpecName: "config-data") pod "216dbcc7-4f75-40eb-9c86-bf4ffff1879f" (UID: "216dbcc7-4f75-40eb-9c86-bf4ffff1879f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.002429 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "216dbcc7-4f75-40eb-9c86-bf4ffff1879f" (UID: "216dbcc7-4f75-40eb-9c86-bf4ffff1879f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.016861 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "216dbcc7-4f75-40eb-9c86-bf4ffff1879f" (UID: "216dbcc7-4f75-40eb-9c86-bf4ffff1879f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.062705 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.063026 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.063116 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6g9q\" (UniqueName: \"kubernetes.io/projected/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-kube-api-access-x6g9q\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.063207 4847 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.063331 4847 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/216dbcc7-4f75-40eb-9c86-bf4ffff1879f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.093483 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.192228 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8b228613-6a71-4c3b-a95a-e0a496c0b346","Type":"ContainerStarted","Data":"e336d759f3c7455c92861bb59948a5b90599d0ed1133379d7356f2b28b492196"} Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.194637 4847 generic.go:334] "Generic (PLEG): container finished" podID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerID="204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733" exitCode=0 Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.194695 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerDied","Data":"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733"} Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.194699 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.194787 4847 scope.go:117] "RemoveContainer" containerID="204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.194773 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"216dbcc7-4f75-40eb-9c86-bf4ffff1879f","Type":"ContainerDied","Data":"cceade013397a1492a8ddcc3b0250468f69711dd8f90ce0e0e33e280caca8f41"} Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.223700 4847 scope.go:117] "RemoveContainer" containerID="d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.228797 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.239302 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.259763 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:55 crc kubenswrapper[4847]: E1210 14:45:55.260257 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.260276 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" Dec 10 14:45:55 crc kubenswrapper[4847]: E1210 14:45:55.260631 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.260652 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.260902 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-metadata" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.260936 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" containerName="nova-metadata-log" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.262666 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.266062 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.266352 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.285429 4847 scope.go:117] "RemoveContainer" containerID="204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.286208 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:55 crc kubenswrapper[4847]: E1210 14:45:55.295039 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733\": container with ID starting with 204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733 not found: ID does not exist" containerID="204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.295099 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733"} err="failed to get container status \"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733\": rpc error: code = NotFound desc = could not find container \"204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733\": container with ID starting with 204f6fbaf3dda008005b67ef67bc00587b9803a8887ae23fd2acc66212587733 not found: ID does not exist" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.295136 4847 scope.go:117] "RemoveContainer" containerID="d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231" Dec 10 14:45:55 crc kubenswrapper[4847]: E1210 14:45:55.295980 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231\": container with ID starting with d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231 not found: ID does not exist" containerID="d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.296029 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231"} err="failed to get container status \"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231\": rpc error: code = NotFound desc = could not find container \"d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231\": container with ID starting with d6a462e2c247eabd8b89761dbec1aa4ccd76dd087b1f839fb2a96d277b177231 not found: ID does not exist" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.368409 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-config-data\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.368524 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f8698b-dd55-4811-81da-55dac46eeffe-logs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.368594 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp6dx\" (UniqueName: \"kubernetes.io/projected/d7f8698b-dd55-4811-81da-55dac46eeffe-kube-api-access-tp6dx\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.368627 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.368700 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.470847 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f8698b-dd55-4811-81da-55dac46eeffe-logs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.470958 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp6dx\" (UniqueName: \"kubernetes.io/projected/d7f8698b-dd55-4811-81da-55dac46eeffe-kube-api-access-tp6dx\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.470997 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.471219 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.471256 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-config-data\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.471393 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7f8698b-dd55-4811-81da-55dac46eeffe-logs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.478297 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.484890 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.491040 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7f8698b-dd55-4811-81da-55dac46eeffe-config-data\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.492353 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp6dx\" (UniqueName: \"kubernetes.io/projected/d7f8698b-dd55-4811-81da-55dac46eeffe-kube-api-access-tp6dx\") pod \"nova-metadata-0\" (UID: \"d7f8698b-dd55-4811-81da-55dac46eeffe\") " pod="openstack/nova-metadata-0" Dec 10 14:45:55 crc kubenswrapper[4847]: I1210 14:45:55.609545 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 14:45:56 crc kubenswrapper[4847]: I1210 14:45:56.063236 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 14:45:56 crc kubenswrapper[4847]: I1210 14:45:56.206452 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7f8698b-dd55-4811-81da-55dac46eeffe","Type":"ContainerStarted","Data":"0359b4cfa527bcf4212ff74920466e038018df74b88d60fd6bb20dd8d4418328"} Dec 10 14:45:56 crc kubenswrapper[4847]: I1210 14:45:56.210194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8b228613-6a71-4c3b-a95a-e0a496c0b346","Type":"ContainerStarted","Data":"0bc3608e5d6ee131d2f2ea0cdf33853d2637e0f6a181414cc5ce41dd9961333b"} Dec 10 14:45:56 crc kubenswrapper[4847]: I1210 14:45:56.232307 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.232291897 podStartE2EDuration="2.232291897s" podCreationTimestamp="2025-12-10 14:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:56.231701382 +0000 UTC m=+1305.800919032" watchObservedRunningTime="2025-12-10 14:45:56.232291897 +0000 UTC m=+1305.801509527" Dec 10 14:45:56 crc kubenswrapper[4847]: I1210 14:45:56.773172 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="216dbcc7-4f75-40eb-9c86-bf4ffff1879f" path="/var/lib/kubelet/pods/216dbcc7-4f75-40eb-9c86-bf4ffff1879f/volumes" Dec 10 14:45:57 crc kubenswrapper[4847]: I1210 14:45:57.224066 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7f8698b-dd55-4811-81da-55dac46eeffe","Type":"ContainerStarted","Data":"981c00df9d449f01c316d9bb6562619c60be26836c303f157507eaa1679ead12"} Dec 10 14:45:57 crc kubenswrapper[4847]: I1210 14:45:57.224748 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d7f8698b-dd55-4811-81da-55dac46eeffe","Type":"ContainerStarted","Data":"bded9ded8c4127cd82b383f8e7a3f96cddd3c8e5812bed994ab73ca1bfb2be96"} Dec 10 14:45:57 crc kubenswrapper[4847]: I1210 14:45:57.252853 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.252828497 podStartE2EDuration="2.252828497s" podCreationTimestamp="2025-12-10 14:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:45:57.24293933 +0000 UTC m=+1306.812156960" watchObservedRunningTime="2025-12-10 14:45:57.252828497 +0000 UTC m=+1306.822046127" Dec 10 14:45:59 crc kubenswrapper[4847]: I1210 14:45:59.605084 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 14:46:00 crc kubenswrapper[4847]: I1210 14:46:00.609659 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 14:46:00 crc kubenswrapper[4847]: I1210 14:46:00.609750 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 14:46:01 crc kubenswrapper[4847]: I1210 14:46:01.010975 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:46:01 crc kubenswrapper[4847]: I1210 14:46:01.011039 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:46:02 crc kubenswrapper[4847]: I1210 14:46:02.828068 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:46:02 crc kubenswrapper[4847]: I1210 14:46:02.828744 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 14:46:03 crc kubenswrapper[4847]: I1210 14:46:03.843860 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0205d3f1-fa73-4291-a22b-1bdd4aaa23ec" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 14:46:03 crc kubenswrapper[4847]: I1210 14:46:03.843918 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0205d3f1-fa73-4291-a22b-1bdd4aaa23ec" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 14:46:04 crc kubenswrapper[4847]: I1210 14:46:04.604281 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 14:46:04 crc kubenswrapper[4847]: I1210 14:46:04.631656 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 14:46:05 crc kubenswrapper[4847]: I1210 14:46:05.329654 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 14:46:05 crc kubenswrapper[4847]: I1210 14:46:05.609936 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 14:46:05 crc kubenswrapper[4847]: I1210 14:46:05.610060 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 14:46:06 crc kubenswrapper[4847]: I1210 14:46:06.621955 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d7f8698b-dd55-4811-81da-55dac46eeffe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 14:46:06 crc kubenswrapper[4847]: I1210 14:46:06.621973 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d7f8698b-dd55-4811-81da-55dac46eeffe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 14:46:10 crc kubenswrapper[4847]: I1210 14:46:10.378906 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 14:46:12 crc kubenswrapper[4847]: I1210 14:46:12.836647 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 14:46:12 crc kubenswrapper[4847]: I1210 14:46:12.837547 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 14:46:12 crc kubenswrapper[4847]: I1210 14:46:12.838539 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 14:46:12 crc kubenswrapper[4847]: I1210 14:46:12.843273 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 14:46:13 crc kubenswrapper[4847]: I1210 14:46:13.378343 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 14:46:13 crc kubenswrapper[4847]: I1210 14:46:13.384437 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 14:46:15 crc kubenswrapper[4847]: I1210 14:46:15.618401 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 14:46:15 crc kubenswrapper[4847]: I1210 14:46:15.618906 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 14:46:15 crc kubenswrapper[4847]: I1210 14:46:15.624231 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 14:46:15 crc kubenswrapper[4847]: I1210 14:46:15.624372 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 14:46:23 crc kubenswrapper[4847]: I1210 14:46:23.469149 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:24 crc kubenswrapper[4847]: I1210 14:46:24.438361 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:28 crc kubenswrapper[4847]: I1210 14:46:28.281041 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="rabbitmq" containerID="cri-o://e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d" gracePeriod=604796 Dec 10 14:46:28 crc kubenswrapper[4847]: I1210 14:46:28.917452 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="rabbitmq" containerID="cri-o://9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683" gracePeriod=604796 Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.012111 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.012511 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.012566 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.013538 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.013608 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187" gracePeriod=600 Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.538257 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187" exitCode=0 Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.538372 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187"} Dec 10 14:46:31 crc kubenswrapper[4847]: I1210 14:46:31.538488 4847 scope.go:117] "RemoveContainer" containerID="173773d3af551dc6ba9a8d3d03e269d8f3affb029b2b3505357043d673d4b2b7" Dec 10 14:46:32 crc kubenswrapper[4847]: I1210 14:46:32.243878 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 10 14:46:32 crc kubenswrapper[4847]: I1210 14:46:32.294869 4847 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 10 14:46:32 crc kubenswrapper[4847]: I1210 14:46:32.550215 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e"} Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.389734 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.537864 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.553794 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.553984 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvb2r\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554049 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554093 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554125 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554154 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554306 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554342 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554389 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554486 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.554535 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.558383 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.559061 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.560826 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.561645 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r" (OuterVolumeSpecName: "kube-api-access-jvb2r") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "kube-api-access-jvb2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.563950 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.565286 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info" (OuterVolumeSpecName: "pod-info") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.588096 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.595089 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.612098 4847 generic.go:334] "Generic (PLEG): container finished" podID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerID="e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d" exitCode=0 Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.612284 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.613086 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerDied","Data":"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d"} Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.613124 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9","Type":"ContainerDied","Data":"7e012cfc4d4b7c04524b283e49bb5fa18e47a46ea2b1517feb4553e55dac581b"} Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.613145 4847 scope.go:117] "RemoveContainer" containerID="e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.621355 4847 generic.go:334] "Generic (PLEG): container finished" podID="2d33f956-6feb-4c03-88f9-95185d25f781" containerID="9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683" exitCode=0 Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.621411 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerDied","Data":"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683"} Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.621440 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2d33f956-6feb-4c03-88f9-95185d25f781","Type":"ContainerDied","Data":"8bcf30f8b59d24440c4be23ee88e7d90cb412343081859d7fb1fbd6b4f4a58d2"} Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.621554 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.655743 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data" (OuterVolumeSpecName: "config-data") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656632 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656691 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9vp2\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656740 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656767 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656796 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656842 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656873 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") pod \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\" (UID: \"7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656912 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.656998 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.657024 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.657093 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.657142 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info\") pod \"2d33f956-6feb-4c03-88f9-95185d25f781\" (UID: \"2d33f956-6feb-4c03-88f9-95185d25f781\") " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.657573 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: W1210 14:46:35.657966 4847 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9/volumes/kubernetes.io~configmap/config-data Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658111 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data" (OuterVolumeSpecName: "config-data") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658465 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf" (OuterVolumeSpecName: "server-conf") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658750 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvb2r\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-kube-api-access-jvb2r\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658778 4847 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658795 4847 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658807 4847 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658837 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658852 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658864 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658876 4847 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658890 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658902 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.658914 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.661347 4847 scope.go:117] "RemoveContainer" containerID="415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.662998 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.663391 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info" (OuterVolumeSpecName: "pod-info") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.664474 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.664655 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.669076 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.674732 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.702999 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2" (OuterVolumeSpecName: "kube-api-access-w9vp2") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "kube-api-access-w9vp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.707825 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.715399 4847 scope.go:117] "RemoveContainer" containerID="e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d" Dec 10 14:46:35 crc kubenswrapper[4847]: E1210 14:46:35.715960 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d\": container with ID starting with e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d not found: ID does not exist" containerID="e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.716094 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d"} err="failed to get container status \"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d\": rpc error: code = NotFound desc = could not find container \"e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d\": container with ID starting with e01ed0ddb42eace3accf78dd48766c06b33f6198c774f517e8391c869f8dbb7d not found: ID does not exist" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.716229 4847 scope.go:117] "RemoveContainer" containerID="415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16" Dec 10 14:46:35 crc kubenswrapper[4847]: E1210 14:46:35.716570 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16\": container with ID starting with 415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16 not found: ID does not exist" containerID="415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.722181 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16"} err="failed to get container status \"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16\": rpc error: code = NotFound desc = could not find container \"415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16\": container with ID starting with 415d78679a6018f50579a473702fe91d0d7f60e5d06183b4428519fd71a8eb16 not found: ID does not exist" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.722434 4847 scope.go:117] "RemoveContainer" containerID="9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.745222 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data" (OuterVolumeSpecName: "config-data") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760194 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760223 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9vp2\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-kube-api-access-w9vp2\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760234 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760243 4847 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2d33f956-6feb-4c03-88f9-95185d25f781-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760271 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760280 4847 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760289 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760297 4847 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2d33f956-6feb-4c03-88f9-95185d25f781-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.760305 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.773999 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf" (OuterVolumeSpecName: "server-conf") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.789872 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.812025 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" (UID: "7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.849866 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2d33f956-6feb-4c03-88f9-95185d25f781" (UID: "2d33f956-6feb-4c03-88f9-95185d25f781"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.861821 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.863694 4847 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2d33f956-6feb-4c03-88f9-95185d25f781-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.871875 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2d33f956-6feb-4c03-88f9-95185d25f781-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.871915 4847 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:35 crc kubenswrapper[4847]: I1210 14:46:35.992515 4847 scope.go:117] "RemoveContainer" containerID="a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.047866 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.058362 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.087635 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.102531 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.103160 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.103261 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.103321 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="setup-container" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.103371 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="setup-container" Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.103448 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.103505 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.103564 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="setup-container" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.103702 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="setup-container" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.103993 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.104066 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" containerName="rabbitmq" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.105179 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.109468 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.110074 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-bbhz4" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.110459 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.110691 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.110913 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.110920 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.114101 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.125802 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.150266 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.160811 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.162942 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.170300 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.170517 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.170618 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.170326 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.170854 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.171040 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.171266 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wvgrd" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.196151 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.233159 4847 scope.go:117] "RemoveContainer" containerID="9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683" Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.233990 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683\": container with ID starting with 9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683 not found: ID does not exist" containerID="9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.234076 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683"} err="failed to get container status \"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683\": rpc error: code = NotFound desc = could not find container \"9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683\": container with ID starting with 9ce1e9a903a166f16fa37dcf04e5e404e80de9feb8405dda74a2888a33942683 not found: ID does not exist" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.234112 4847 scope.go:117] "RemoveContainer" containerID="a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05" Dec 10 14:46:36 crc kubenswrapper[4847]: E1210 14:46:36.234855 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05\": container with ID starting with a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05 not found: ID does not exist" containerID="a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.234880 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05"} err="failed to get container status \"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05\": rpc error: code = NotFound desc = could not find container \"a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05\": container with ID starting with a329b0b34b6ff2760e9c20a1675ca5097e207e4558045a7ad2a4b37ba101ec05 not found: ID does not exist" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.282952 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283023 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-config-data\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283082 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xj5\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-kube-api-access-j5xj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283115 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283231 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283304 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283334 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75428316-3c6c-4166-a55e-cb293dbad536-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283481 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283597 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283653 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd9zj\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-kube-api-access-hd9zj\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283687 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283792 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283864 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283916 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283943 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.283989 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284015 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75428316-3c6c-4166-a55e-cb293dbad536-pod-info\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284036 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-server-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284114 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284148 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284177 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.284211 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386365 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386420 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386447 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386466 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75428316-3c6c-4166-a55e-cb293dbad536-pod-info\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386488 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-server-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386506 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386535 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386558 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386582 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386607 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386637 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386660 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-config-data\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386689 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xj5\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-kube-api-access-j5xj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386717 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386771 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386818 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386842 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75428316-3c6c-4166-a55e-cb293dbad536-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386883 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386930 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386952 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd9zj\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-kube-api-access-hd9zj\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.386973 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.387009 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.388935 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.389457 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.389646 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-config-data\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.389887 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.390462 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.390524 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.390638 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.390800 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.392361 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.392617 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.392756 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.393593 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75428316-3c6c-4166-a55e-cb293dbad536-server-conf\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.401791 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.406533 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xj5\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-kube-api-access-j5xj5\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.407199 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.410255 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75428316-3c6c-4166-a55e-cb293dbad536-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.410410 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75428316-3c6c-4166-a55e-cb293dbad536-pod-info\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.410991 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.410985 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.411572 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd9zj\" (UniqueName: \"kubernetes.io/projected/75428316-3c6c-4166-a55e-cb293dbad536-kube-api-access-hd9zj\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.421896 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.422164 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.438329 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.447067 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"75428316-3c6c-4166-a55e-cb293dbad536\") " pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.506601 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.752171 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.774945 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d33f956-6feb-4c03-88f9-95185d25f781" path="/var/lib/kubelet/pods/2d33f956-6feb-4c03-88f9-95185d25f781/volumes" Dec 10 14:46:36 crc kubenswrapper[4847]: I1210 14:46:36.776677 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9" path="/var/lib/kubelet/pods/7f5ccab8-d00d-43e4-938c-e49a1cdfc3d9/volumes" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.051764 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.261975 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 14:46:37 crc kubenswrapper[4847]: W1210 14:46:37.266856 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75428316_3c6c_4166_a55e_cb293dbad536.slice/crio-2d0757802b6339bee3c7bc4b2445d50e24e5b50a0b0e706d0392eb346740415f WatchSource:0}: Error finding container 2d0757802b6339bee3c7bc4b2445d50e24e5b50a0b0e706d0392eb346740415f: Status 404 returned error can't find the container with id 2d0757802b6339bee3c7bc4b2445d50e24e5b50a0b0e706d0392eb346740415f Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.687655 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"75428316-3c6c-4166-a55e-cb293dbad536","Type":"ContainerStarted","Data":"2d0757802b6339bee3c7bc4b2445d50e24e5b50a0b0e706d0392eb346740415f"} Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.689246 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf","Type":"ContainerStarted","Data":"adaadeb47e6f557990f012560f2de760457f5fe7e167a41e05c951899f65d725"} Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.866396 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9p6ms"] Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.868298 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.891034 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.897218 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9p6ms"] Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.930881 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.930940 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.930966 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.930993 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.931017 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.931111 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkpk\" (UniqueName: \"kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:37 crc kubenswrapper[4847]: I1210 14:46:37.931141 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.012994 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9p6ms"] Dec 10 14:46:38 crc kubenswrapper[4847]: E1210 14:46:38.014353 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-rkkpk openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-d558885bc-9p6ms" podUID="3117a764-5c7d-46a0-9b94-308db2944ff7" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032737 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkpk\" (UniqueName: \"kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032792 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032883 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032926 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032961 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.032997 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.033033 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.033931 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.033934 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.034188 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.034268 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.034444 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.034966 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.049830 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mbxst"] Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.051471 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.073261 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkpk\" (UniqueName: \"kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk\") pod \"dnsmasq-dns-d558885bc-9p6ms\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.085230 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mbxst"] Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.136094 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.136192 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.136308 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdfjg\" (UniqueName: \"kubernetes.io/projected/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-kube-api-access-rdfjg\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.136391 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.136575 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.137100 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.137290 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-config\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239470 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239546 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239690 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239760 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-config\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239790 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239810 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.239879 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdfjg\" (UniqueName: \"kubernetes.io/projected/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-kube-api-access-rdfjg\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.240968 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.241284 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.241880 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-config\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.241936 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.242325 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.243137 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.257871 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdfjg\" (UniqueName: \"kubernetes.io/projected/70f2cfd7-1fd9-48f5-8004-6df8f4d428c9-kube-api-access-rdfjg\") pod \"dnsmasq-dns-78c64bc9c5-mbxst\" (UID: \"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9\") " pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.429449 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.717022 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.717078 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf","Type":"ContainerStarted","Data":"ef2181764111d6cd2272ebbe770783038c5807a6897fd491290eac8402aecd91"} Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.770475 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.858649 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.858894 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.858935 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.858981 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkkpk\" (UniqueName: \"kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.859087 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.859127 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.859157 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam\") pod \"3117a764-5c7d-46a0-9b94-308db2944ff7\" (UID: \"3117a764-5c7d-46a0-9b94-308db2944ff7\") " Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.860868 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.861391 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.862392 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.868536 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.870287 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.870420 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config" (OuterVolumeSpecName: "config") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.878017 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk" (OuterVolumeSpecName: "kube-api-access-rkkpk") pod "3117a764-5c7d-46a0-9b94-308db2944ff7" (UID: "3117a764-5c7d-46a0-9b94-308db2944ff7"). InnerVolumeSpecName "kube-api-access-rkkpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962068 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962113 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962122 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkkpk\" (UniqueName: \"kubernetes.io/projected/3117a764-5c7d-46a0-9b94-308db2944ff7-kube-api-access-rkkpk\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962134 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962143 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962150 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:38 crc kubenswrapper[4847]: I1210 14:46:38.962158 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3117a764-5c7d-46a0-9b94-308db2944ff7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.180965 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-mbxst"] Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.730194 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" event={"ID":"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9","Type":"ContainerStarted","Data":"2f3466d4a573a1e3a495c33fa35e8e18a11bfbc2ccd924486572b93d7c4e7588"} Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.732092 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-9p6ms" Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.732906 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"75428316-3c6c-4166-a55e-cb293dbad536","Type":"ContainerStarted","Data":"c4f11eaf306629a061e87404a5f639312cef12fe5050eb39896cf72c06e10505"} Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.818160 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9p6ms"] Dec 10 14:46:39 crc kubenswrapper[4847]: I1210 14:46:39.827529 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-9p6ms"] Dec 10 14:46:40 crc kubenswrapper[4847]: I1210 14:46:40.742736 4847 generic.go:334] "Generic (PLEG): container finished" podID="70f2cfd7-1fd9-48f5-8004-6df8f4d428c9" containerID="cac7f765f89c296cec7a443e959cdc566777633937575ebe18d2781027a31d46" exitCode=0 Dec 10 14:46:40 crc kubenswrapper[4847]: I1210 14:46:40.742822 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" event={"ID":"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9","Type":"ContainerDied","Data":"cac7f765f89c296cec7a443e959cdc566777633937575ebe18d2781027a31d46"} Dec 10 14:46:40 crc kubenswrapper[4847]: I1210 14:46:40.786286 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3117a764-5c7d-46a0-9b94-308db2944ff7" path="/var/lib/kubelet/pods/3117a764-5c7d-46a0-9b94-308db2944ff7/volumes" Dec 10 14:46:42 crc kubenswrapper[4847]: I1210 14:46:42.769655 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" event={"ID":"70f2cfd7-1fd9-48f5-8004-6df8f4d428c9","Type":"ContainerStarted","Data":"2f353cbae1ae0a5e3edf3f433f9159a16dfe1c096d3464337e5ca4dd2a809c97"} Dec 10 14:46:45 crc kubenswrapper[4847]: I1210 14:46:45.798663 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:45 crc kubenswrapper[4847]: I1210 14:46:45.826821 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" podStartSLOduration=7.826801781 podStartE2EDuration="7.826801781s" podCreationTimestamp="2025-12-10 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:46:45.822019836 +0000 UTC m=+1355.391237476" watchObservedRunningTime="2025-12-10 14:46:45.826801781 +0000 UTC m=+1355.396019411" Dec 10 14:46:46 crc kubenswrapper[4847]: I1210 14:46:46.807871 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-mbxst" Dec 10 14:46:46 crc kubenswrapper[4847]: I1210 14:46:46.873168 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:46:46 crc kubenswrapper[4847]: I1210 14:46:46.873930 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="dnsmasq-dns" containerID="cri-o://f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36" gracePeriod=10 Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.449141 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.647857 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.648064 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.648105 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.648137 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.648213 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.648254 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfsg4\" (UniqueName: \"kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4\") pod \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\" (UID: \"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb\") " Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.654009 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4" (OuterVolumeSpecName: "kube-api-access-qfsg4") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "kube-api-access-qfsg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.701695 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.704040 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config" (OuterVolumeSpecName: "config") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.704330 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.705073 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.706056 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" (UID: "ff38ca48-f4db-4ad4-9a86-ae3c70b669fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.751934 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.751969 4847 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.751977 4847 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.751985 4847 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.751995 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfsg4\" (UniqueName: \"kubernetes.io/projected/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-kube-api-access-qfsg4\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.752004 4847 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb-config\") on node \"crc\" DevicePath \"\"" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.815741 4847 generic.go:334] "Generic (PLEG): container finished" podID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerID="f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36" exitCode=0 Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.815798 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.815835 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" event={"ID":"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb","Type":"ContainerDied","Data":"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36"} Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.815862 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dd2sj" event={"ID":"ff38ca48-f4db-4ad4-9a86-ae3c70b669fb","Type":"ContainerDied","Data":"adf77893fad6ccba89507973cb11cd6be97170c067f79acb132aafc4f2df7cf8"} Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.815878 4847 scope.go:117] "RemoveContainer" containerID="f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.853708 4847 scope.go:117] "RemoveContainer" containerID="066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.870557 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.891852 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dd2sj"] Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.893348 4847 scope.go:117] "RemoveContainer" containerID="f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36" Dec 10 14:46:47 crc kubenswrapper[4847]: E1210 14:46:47.893818 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36\": container with ID starting with f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36 not found: ID does not exist" containerID="f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.893856 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36"} err="failed to get container status \"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36\": rpc error: code = NotFound desc = could not find container \"f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36\": container with ID starting with f38124a9e5b86704e30f6676e856ca396ddcb681f320d4095fd40f6bdf944c36 not found: ID does not exist" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.893884 4847 scope.go:117] "RemoveContainer" containerID="066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9" Dec 10 14:46:47 crc kubenswrapper[4847]: E1210 14:46:47.894124 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9\": container with ID starting with 066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9 not found: ID does not exist" containerID="066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9" Dec 10 14:46:47 crc kubenswrapper[4847]: I1210 14:46:47.894148 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9"} err="failed to get container status \"066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9\": rpc error: code = NotFound desc = could not find container \"066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9\": container with ID starting with 066591dce49a1bfe8b3013db0e30f945e500935e0e2a6d09ca832764dce1ffe9 not found: ID does not exist" Dec 10 14:46:48 crc kubenswrapper[4847]: I1210 14:46:48.769639 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" path="/var/lib/kubelet/pods/ff38ca48-f4db-4ad4-9a86-ae3c70b669fb/volumes" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.033290 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd"] Dec 10 14:46:57 crc kubenswrapper[4847]: E1210 14:46:57.034556 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="init" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.034573 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="init" Dec 10 14:46:57 crc kubenswrapper[4847]: E1210 14:46:57.034599 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="dnsmasq-dns" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.034607 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="dnsmasq-dns" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.034919 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff38ca48-f4db-4ad4-9a86-ae3c70b669fb" containerName="dnsmasq-dns" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.035647 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.041231 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.041288 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.041336 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.042356 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.046339 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd"] Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.138933 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tvxg\" (UniqueName: \"kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.139170 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.139212 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.139260 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.241808 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tvxg\" (UniqueName: \"kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.242005 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.242039 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.242125 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.249534 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.249973 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.252377 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.266477 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tvxg\" (UniqueName: \"kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.362994 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.978927 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd"] Dec 10 14:46:57 crc kubenswrapper[4847]: W1210 14:46:57.987134 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f79a99b_09b0_4951_bf05_42fe68d44e53.slice/crio-fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7 WatchSource:0}: Error finding container fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7: Status 404 returned error can't find the container with id fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7 Dec 10 14:46:57 crc kubenswrapper[4847]: I1210 14:46:57.990451 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:46:58 crc kubenswrapper[4847]: I1210 14:46:58.934935 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" event={"ID":"1f79a99b-09b0-4951-bf05-42fe68d44e53","Type":"ContainerStarted","Data":"fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7"} Dec 10 14:47:08 crc kubenswrapper[4847]: I1210 14:47:08.047219 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" event={"ID":"1f79a99b-09b0-4951-bf05-42fe68d44e53","Type":"ContainerStarted","Data":"171366fd1b3fc8960af94882d36c203129373ca3d0bdffd2e5a84353d3e957a8"} Dec 10 14:47:08 crc kubenswrapper[4847]: I1210 14:47:08.066048 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" podStartSLOduration=1.61849298 podStartE2EDuration="11.066018483s" podCreationTimestamp="2025-12-10 14:46:57 +0000 UTC" firstStartedPulling="2025-12-10 14:46:57.99017034 +0000 UTC m=+1367.559387980" lastFinishedPulling="2025-12-10 14:47:07.437695843 +0000 UTC m=+1377.006913483" observedRunningTime="2025-12-10 14:47:08.064405709 +0000 UTC m=+1377.633623329" watchObservedRunningTime="2025-12-10 14:47:08.066018483 +0000 UTC m=+1377.635236113" Dec 10 14:47:11 crc kubenswrapper[4847]: I1210 14:47:11.098945 4847 generic.go:334] "Generic (PLEG): container finished" podID="75428316-3c6c-4166-a55e-cb293dbad536" containerID="c4f11eaf306629a061e87404a5f639312cef12fe5050eb39896cf72c06e10505" exitCode=0 Dec 10 14:47:11 crc kubenswrapper[4847]: I1210 14:47:11.099046 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"75428316-3c6c-4166-a55e-cb293dbad536","Type":"ContainerDied","Data":"c4f11eaf306629a061e87404a5f639312cef12fe5050eb39896cf72c06e10505"} Dec 10 14:47:11 crc kubenswrapper[4847]: I1210 14:47:11.101581 4847 generic.go:334] "Generic (PLEG): container finished" podID="0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf" containerID="ef2181764111d6cd2272ebbe770783038c5807a6897fd491290eac8402aecd91" exitCode=0 Dec 10 14:47:11 crc kubenswrapper[4847]: I1210 14:47:11.101606 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf","Type":"ContainerDied","Data":"ef2181764111d6cd2272ebbe770783038c5807a6897fd491290eac8402aecd91"} Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.113225 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"75428316-3c6c-4166-a55e-cb293dbad536","Type":"ContainerStarted","Data":"fb605f40e7c2ebbec2d3d40b90a8a5ea2e8cf616cb7776da2b5db41da4afe72d"} Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.114090 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.115523 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf","Type":"ContainerStarted","Data":"7874d3537519c6014bcc9d1706dea3a2b975b97ff9fb04f5862cb004b204c8ce"} Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.116817 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.145669 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.1456513 podStartE2EDuration="36.1456513s" podCreationTimestamp="2025-12-10 14:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:47:12.144324514 +0000 UTC m=+1381.713542164" watchObservedRunningTime="2025-12-10 14:47:12.1456513 +0000 UTC m=+1381.714868930" Dec 10 14:47:12 crc kubenswrapper[4847]: I1210 14:47:12.183406 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.183386728 podStartE2EDuration="36.183386728s" podCreationTimestamp="2025-12-10 14:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 14:47:12.177070906 +0000 UTC m=+1381.746288536" watchObservedRunningTime="2025-12-10 14:47:12.183386728 +0000 UTC m=+1381.752604358" Dec 10 14:47:19 crc kubenswrapper[4847]: I1210 14:47:19.192379 4847 generic.go:334] "Generic (PLEG): container finished" podID="1f79a99b-09b0-4951-bf05-42fe68d44e53" containerID="171366fd1b3fc8960af94882d36c203129373ca3d0bdffd2e5a84353d3e957a8" exitCode=0 Dec 10 14:47:19 crc kubenswrapper[4847]: I1210 14:47:19.192502 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" event={"ID":"1f79a99b-09b0-4951-bf05-42fe68d44e53","Type":"ContainerDied","Data":"171366fd1b3fc8960af94882d36c203129373ca3d0bdffd2e5a84353d3e957a8"} Dec 10 14:47:20 crc kubenswrapper[4847]: I1210 14:47:20.456160 4847 scope.go:117] "RemoveContainer" containerID="5cabca541076c892b3958799bd20191127fce0fc669d98986f7e091b71b70b59" Dec 10 14:47:20 crc kubenswrapper[4847]: I1210 14:47:20.978864 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.093180 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tvxg\" (UniqueName: \"kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg\") pod \"1f79a99b-09b0-4951-bf05-42fe68d44e53\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.093603 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key\") pod \"1f79a99b-09b0-4951-bf05-42fe68d44e53\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.093796 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory\") pod \"1f79a99b-09b0-4951-bf05-42fe68d44e53\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.093946 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle\") pod \"1f79a99b-09b0-4951-bf05-42fe68d44e53\" (UID: \"1f79a99b-09b0-4951-bf05-42fe68d44e53\") " Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.100925 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1f79a99b-09b0-4951-bf05-42fe68d44e53" (UID: "1f79a99b-09b0-4951-bf05-42fe68d44e53"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.102330 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg" (OuterVolumeSpecName: "kube-api-access-2tvxg") pod "1f79a99b-09b0-4951-bf05-42fe68d44e53" (UID: "1f79a99b-09b0-4951-bf05-42fe68d44e53"). InnerVolumeSpecName "kube-api-access-2tvxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.127038 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1f79a99b-09b0-4951-bf05-42fe68d44e53" (UID: "1f79a99b-09b0-4951-bf05-42fe68d44e53"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.130870 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory" (OuterVolumeSpecName: "inventory") pod "1f79a99b-09b0-4951-bf05-42fe68d44e53" (UID: "1f79a99b-09b0-4951-bf05-42fe68d44e53"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.197351 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.197428 4847 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.197450 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tvxg\" (UniqueName: \"kubernetes.io/projected/1f79a99b-09b0-4951-bf05-42fe68d44e53-kube-api-access-2tvxg\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.197461 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f79a99b-09b0-4951-bf05-42fe68d44e53-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.217976 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" event={"ID":"1f79a99b-09b0-4951-bf05-42fe68d44e53","Type":"ContainerDied","Data":"fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7"} Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.218347 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe355c2c85e5f714e58e028fd79981270bc487962b6a913f47fd53f926bb2b7" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.218256 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.364668 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz"] Dec 10 14:47:21 crc kubenswrapper[4847]: E1210 14:47:21.365353 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f79a99b-09b0-4951-bf05-42fe68d44e53" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.365379 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f79a99b-09b0-4951-bf05-42fe68d44e53" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.365688 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f79a99b-09b0-4951-bf05-42fe68d44e53" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.366734 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.371059 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.371330 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.371539 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.380175 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz"] Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.380315 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.420288 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrzzl\" (UniqueName: \"kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.420367 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.420460 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.522556 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrzzl\" (UniqueName: \"kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.522635 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.522709 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.527248 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.538105 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.559298 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrzzl\" (UniqueName: \"kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s87nz\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:21 crc kubenswrapper[4847]: I1210 14:47:21.685980 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:22 crc kubenswrapper[4847]: I1210 14:47:22.312135 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz"] Dec 10 14:47:23 crc kubenswrapper[4847]: I1210 14:47:23.242397 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" event={"ID":"d39485c7-eaab-4829-b978-ca8109f82227","Type":"ContainerStarted","Data":"f260de3e7a5a19035e981b923a399cf13a1e23012b42d23810f2f680a30fb934"} Dec 10 14:47:23 crc kubenswrapper[4847]: I1210 14:47:23.242805 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" event={"ID":"d39485c7-eaab-4829-b978-ca8109f82227","Type":"ContainerStarted","Data":"a1e7705260f6b32655f88ccd1c912e9722be51512b52bc6fd7b14417cbc87a42"} Dec 10 14:47:23 crc kubenswrapper[4847]: I1210 14:47:23.304634 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" podStartSLOduration=1.753280063 podStartE2EDuration="2.304612995s" podCreationTimestamp="2025-12-10 14:47:21 +0000 UTC" firstStartedPulling="2025-12-10 14:47:22.314186709 +0000 UTC m=+1391.883404339" lastFinishedPulling="2025-12-10 14:47:22.865519611 +0000 UTC m=+1392.434737271" observedRunningTime="2025-12-10 14:47:23.29267098 +0000 UTC m=+1392.861888610" watchObservedRunningTime="2025-12-10 14:47:23.304612995 +0000 UTC m=+1392.873830625" Dec 10 14:47:26 crc kubenswrapper[4847]: I1210 14:47:26.271270 4847 generic.go:334] "Generic (PLEG): container finished" podID="d39485c7-eaab-4829-b978-ca8109f82227" containerID="f260de3e7a5a19035e981b923a399cf13a1e23012b42d23810f2f680a30fb934" exitCode=0 Dec 10 14:47:26 crc kubenswrapper[4847]: I1210 14:47:26.271340 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" event={"ID":"d39485c7-eaab-4829-b978-ca8109f82227","Type":"ContainerDied","Data":"f260de3e7a5a19035e981b923a399cf13a1e23012b42d23810f2f680a30fb934"} Dec 10 14:47:26 crc kubenswrapper[4847]: I1210 14:47:26.509244 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 14:47:26 crc kubenswrapper[4847]: I1210 14:47:26.800970 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.746639 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.855440 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key\") pod \"d39485c7-eaab-4829-b978-ca8109f82227\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.855614 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrzzl\" (UniqueName: \"kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl\") pod \"d39485c7-eaab-4829-b978-ca8109f82227\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.855747 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory\") pod \"d39485c7-eaab-4829-b978-ca8109f82227\" (UID: \"d39485c7-eaab-4829-b978-ca8109f82227\") " Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.863777 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl" (OuterVolumeSpecName: "kube-api-access-zrzzl") pod "d39485c7-eaab-4829-b978-ca8109f82227" (UID: "d39485c7-eaab-4829-b978-ca8109f82227"). InnerVolumeSpecName "kube-api-access-zrzzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.898028 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d39485c7-eaab-4829-b978-ca8109f82227" (UID: "d39485c7-eaab-4829-b978-ca8109f82227"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.912353 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory" (OuterVolumeSpecName: "inventory") pod "d39485c7-eaab-4829-b978-ca8109f82227" (UID: "d39485c7-eaab-4829-b978-ca8109f82227"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.959962 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrzzl\" (UniqueName: \"kubernetes.io/projected/d39485c7-eaab-4829-b978-ca8109f82227-kube-api-access-zrzzl\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.960055 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:27 crc kubenswrapper[4847]: I1210 14:47:27.960070 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d39485c7-eaab-4829-b978-ca8109f82227-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.291815 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" event={"ID":"d39485c7-eaab-4829-b978-ca8109f82227","Type":"ContainerDied","Data":"a1e7705260f6b32655f88ccd1c912e9722be51512b52bc6fd7b14417cbc87a42"} Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.291858 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1e7705260f6b32655f88ccd1c912e9722be51512b52bc6fd7b14417cbc87a42" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.291904 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s87nz" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.426525 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm"] Dec 10 14:47:28 crc kubenswrapper[4847]: E1210 14:47:28.427138 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39485c7-eaab-4829-b978-ca8109f82227" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.427235 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39485c7-eaab-4829-b978-ca8109f82227" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.427506 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39485c7-eaab-4829-b978-ca8109f82227" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.428165 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.429970 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.430214 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.430842 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.430888 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.445601 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm"] Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.469612 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gffl4\" (UniqueName: \"kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.469666 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.469727 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.469833 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.571890 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gffl4\" (UniqueName: \"kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.572285 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.572342 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.572425 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.577236 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.577242 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.577451 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.591114 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gffl4\" (UniqueName: \"kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:28 crc kubenswrapper[4847]: I1210 14:47:28.835100 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:47:29 crc kubenswrapper[4847]: I1210 14:47:29.375084 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm"] Dec 10 14:47:30 crc kubenswrapper[4847]: I1210 14:47:30.310845 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" event={"ID":"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6","Type":"ContainerStarted","Data":"7ef3cf3ed8df4801c6249a157fc453aa18140c35aaa37903117854e9f6f70972"} Dec 10 14:47:30 crc kubenswrapper[4847]: I1210 14:47:30.311451 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" event={"ID":"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6","Type":"ContainerStarted","Data":"729e0e896ea151ca6317345285f24f1876f81ae2f21083224decda0b7b3d2cfa"} Dec 10 14:47:30 crc kubenswrapper[4847]: I1210 14:47:30.335606 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" podStartSLOduration=1.8112072590000001 podStartE2EDuration="2.335590557s" podCreationTimestamp="2025-12-10 14:47:28 +0000 UTC" firstStartedPulling="2025-12-10 14:47:29.377259395 +0000 UTC m=+1398.946477015" lastFinishedPulling="2025-12-10 14:47:29.901642683 +0000 UTC m=+1399.470860313" observedRunningTime="2025-12-10 14:47:30.330053626 +0000 UTC m=+1399.899271246" watchObservedRunningTime="2025-12-10 14:47:30.335590557 +0000 UTC m=+1399.904808187" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.296902 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djwns"] Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.302806 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.309177 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djwns"] Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.438061 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72fl\" (UniqueName: \"kubernetes.io/projected/902cf9a3-e8c7-46c9-b058-de3249e6ce65-kube-api-access-r72fl\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.438141 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-utilities\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.438188 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-catalog-content\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.539391 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72fl\" (UniqueName: \"kubernetes.io/projected/902cf9a3-e8c7-46c9-b058-de3249e6ce65-kube-api-access-r72fl\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.539454 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-utilities\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.539506 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-catalog-content\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.540051 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-utilities\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.540257 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/902cf9a3-e8c7-46c9-b058-de3249e6ce65-catalog-content\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.580304 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72fl\" (UniqueName: \"kubernetes.io/projected/902cf9a3-e8c7-46c9-b058-de3249e6ce65-kube-api-access-r72fl\") pod \"redhat-operators-djwns\" (UID: \"902cf9a3-e8c7-46c9-b058-de3249e6ce65\") " pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:01 crc kubenswrapper[4847]: I1210 14:48:01.627584 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:02 crc kubenswrapper[4847]: I1210 14:48:02.096665 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djwns"] Dec 10 14:48:02 crc kubenswrapper[4847]: I1210 14:48:02.606276 4847 generic.go:334] "Generic (PLEG): container finished" podID="902cf9a3-e8c7-46c9-b058-de3249e6ce65" containerID="669b8db3f88f9eab13ba6373e454154b259f53c13b210cdecd6cc8e5fe1a5412" exitCode=0 Dec 10 14:48:02 crc kubenswrapper[4847]: I1210 14:48:02.606664 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djwns" event={"ID":"902cf9a3-e8c7-46c9-b058-de3249e6ce65","Type":"ContainerDied","Data":"669b8db3f88f9eab13ba6373e454154b259f53c13b210cdecd6cc8e5fe1a5412"} Dec 10 14:48:02 crc kubenswrapper[4847]: I1210 14:48:02.606703 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djwns" event={"ID":"902cf9a3-e8c7-46c9-b058-de3249e6ce65","Type":"ContainerStarted","Data":"4116dab5cda62a653d69f2fabc5bc31d8bb8fc717d28c96818f64c90a9e7ac57"} Dec 10 14:48:13 crc kubenswrapper[4847]: I1210 14:48:13.707784 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djwns" event={"ID":"902cf9a3-e8c7-46c9-b058-de3249e6ce65","Type":"ContainerStarted","Data":"e37c5e76b8bbfbc3b80bb7e2e49195be079e9347ee73af4090cd67dbb06e31d1"} Dec 10 14:48:18 crc kubenswrapper[4847]: I1210 14:48:18.757173 4847 generic.go:334] "Generic (PLEG): container finished" podID="902cf9a3-e8c7-46c9-b058-de3249e6ce65" containerID="e37c5e76b8bbfbc3b80bb7e2e49195be079e9347ee73af4090cd67dbb06e31d1" exitCode=0 Dec 10 14:48:18 crc kubenswrapper[4847]: I1210 14:48:18.757244 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djwns" event={"ID":"902cf9a3-e8c7-46c9-b058-de3249e6ce65","Type":"ContainerDied","Data":"e37c5e76b8bbfbc3b80bb7e2e49195be079e9347ee73af4090cd67dbb06e31d1"} Dec 10 14:48:20 crc kubenswrapper[4847]: I1210 14:48:20.780300 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djwns" event={"ID":"902cf9a3-e8c7-46c9-b058-de3249e6ce65","Type":"ContainerStarted","Data":"783a09bb0a7923e2921b879ed402c4c6cdf4644d522b5e3ec85d10e7a07eb7e1"} Dec 10 14:48:20 crc kubenswrapper[4847]: I1210 14:48:20.809282 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djwns" podStartSLOduration=2.8501955 podStartE2EDuration="19.809255731s" podCreationTimestamp="2025-12-10 14:48:01 +0000 UTC" firstStartedPulling="2025-12-10 14:48:02.608816927 +0000 UTC m=+1432.178034557" lastFinishedPulling="2025-12-10 14:48:19.567877158 +0000 UTC m=+1449.137094788" observedRunningTime="2025-12-10 14:48:20.799119565 +0000 UTC m=+1450.368337205" watchObservedRunningTime="2025-12-10 14:48:20.809255731 +0000 UTC m=+1450.378473361" Dec 10 14:48:21 crc kubenswrapper[4847]: I1210 14:48:21.503682 4847 scope.go:117] "RemoveContainer" containerID="a8df61f54bd8405abd20ac66610931007cf65b734093972a9a9ae7133c03d0ea" Dec 10 14:48:21 crc kubenswrapper[4847]: I1210 14:48:21.537310 4847 scope.go:117] "RemoveContainer" containerID="68ea8b37c6643ca1fc7703f6f8985f1cf33536eeab82c06721873fa242986cb4" Dec 10 14:48:21 crc kubenswrapper[4847]: I1210 14:48:21.589362 4847 scope.go:117] "RemoveContainer" containerID="00143c0dae6df68b21422196c533659d71e6ee8f5a37e6ef096448d74f03b88e" Dec 10 14:48:21 crc kubenswrapper[4847]: I1210 14:48:21.627810 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:21 crc kubenswrapper[4847]: I1210 14:48:21.627862 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:22 crc kubenswrapper[4847]: I1210 14:48:22.687462 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-djwns" podUID="902cf9a3-e8c7-46c9-b058-de3249e6ce65" containerName="registry-server" probeResult="failure" output=< Dec 10 14:48:22 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:48:22 crc kubenswrapper[4847]: > Dec 10 14:48:31 crc kubenswrapper[4847]: I1210 14:48:31.011614 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:48:31 crc kubenswrapper[4847]: I1210 14:48:31.012645 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:48:31 crc kubenswrapper[4847]: I1210 14:48:31.676014 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:31 crc kubenswrapper[4847]: I1210 14:48:31.730237 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djwns" Dec 10 14:48:32 crc kubenswrapper[4847]: I1210 14:48:32.327382 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djwns"] Dec 10 14:48:32 crc kubenswrapper[4847]: I1210 14:48:32.499209 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:48:32 crc kubenswrapper[4847]: I1210 14:48:32.499447 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tc2pm" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="registry-server" containerID="cri-o://ac07a3f731a912f09e8666e9cba3737e28d16719a0a6813fce4536c2dd7402f6" gracePeriod=2 Dec 10 14:48:32 crc kubenswrapper[4847]: I1210 14:48:32.908993 4847 generic.go:334] "Generic (PLEG): container finished" podID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerID="ac07a3f731a912f09e8666e9cba3737e28d16719a0a6813fce4536c2dd7402f6" exitCode=0 Dec 10 14:48:32 crc kubenswrapper[4847]: I1210 14:48:32.910122 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerDied","Data":"ac07a3f731a912f09e8666e9cba3737e28d16719a0a6813fce4536c2dd7402f6"} Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.020012 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.116589 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content\") pod \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.117490 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities\") pod \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.117602 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6927\" (UniqueName: \"kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927\") pod \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\" (UID: \"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe\") " Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.117675 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities" (OuterVolumeSpecName: "utilities") pod "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" (UID: "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.118408 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.124249 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927" (OuterVolumeSpecName: "kube-api-access-n6927") pod "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" (UID: "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe"). InnerVolumeSpecName "kube-api-access-n6927". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.221375 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6927\" (UniqueName: \"kubernetes.io/projected/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-kube-api-access-n6927\") on node \"crc\" DevicePath \"\"" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.250767 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" (UID: "dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.324004 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.924737 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tc2pm" event={"ID":"dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe","Type":"ContainerDied","Data":"b9fa17489a2d2fdbbb112a6c7d8550bc1faf33b3bf6afe175656cd312db7c5a0"} Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.924793 4847 scope.go:117] "RemoveContainer" containerID="ac07a3f731a912f09e8666e9cba3737e28d16719a0a6813fce4536c2dd7402f6" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.924871 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tc2pm" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.965483 4847 scope.go:117] "RemoveContainer" containerID="40fe880cefdae4b028349b8568f3cbc966ff7af1c9aafe1b94f84db3994bb30f" Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.966189 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.982459 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tc2pm"] Dec 10 14:48:33 crc kubenswrapper[4847]: I1210 14:48:33.996178 4847 scope.go:117] "RemoveContainer" containerID="2dad778a5c82fabc6cd9c7e488a77aaca4374e37e7f60880ba560c5481da0b4d" Dec 10 14:48:34 crc kubenswrapper[4847]: I1210 14:48:34.772519 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" path="/var/lib/kubelet/pods/dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe/volumes" Dec 10 14:49:01 crc kubenswrapper[4847]: I1210 14:49:01.011192 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:49:01 crc kubenswrapper[4847]: I1210 14:49:01.011738 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.724424 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:08 crc kubenswrapper[4847]: E1210 14:49:08.725358 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="extract-utilities" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.725377 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="extract-utilities" Dec 10 14:49:08 crc kubenswrapper[4847]: E1210 14:49:08.725408 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="registry-server" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.725417 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="registry-server" Dec 10 14:49:08 crc kubenswrapper[4847]: E1210 14:49:08.725450 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="extract-content" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.725458 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="extract-content" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.725640 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc69be8a-98d2-4ebe-a1b7-a2df888c7bbe" containerName="registry-server" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.727107 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.740793 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.803557 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5gm7\" (UniqueName: \"kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.803707 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.803810 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.906140 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5gm7\" (UniqueName: \"kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.906220 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.906255 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.906822 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.906868 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:08 crc kubenswrapper[4847]: I1210 14:49:08.933583 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5gm7\" (UniqueName: \"kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7\") pod \"redhat-marketplace-trb8d\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:09 crc kubenswrapper[4847]: I1210 14:49:09.048114 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:09 crc kubenswrapper[4847]: I1210 14:49:09.589878 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:09 crc kubenswrapper[4847]: W1210 14:49:09.604289 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4df3602_2c3c_4428_99e7_a40ab62cffe6.slice/crio-543167e5dd769eda57b5df35ab3626abd99ae6fcde0daf60d7741f820674595b WatchSource:0}: Error finding container 543167e5dd769eda57b5df35ab3626abd99ae6fcde0daf60d7741f820674595b: Status 404 returned error can't find the container with id 543167e5dd769eda57b5df35ab3626abd99ae6fcde0daf60d7741f820674595b Dec 10 14:49:10 crc kubenswrapper[4847]: I1210 14:49:10.300802 4847 generic.go:334] "Generic (PLEG): container finished" podID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerID="3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094" exitCode=0 Dec 10 14:49:10 crc kubenswrapper[4847]: I1210 14:49:10.300875 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerDied","Data":"3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094"} Dec 10 14:49:10 crc kubenswrapper[4847]: I1210 14:49:10.301226 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerStarted","Data":"543167e5dd769eda57b5df35ab3626abd99ae6fcde0daf60d7741f820674595b"} Dec 10 14:49:11 crc kubenswrapper[4847]: I1210 14:49:11.320638 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerStarted","Data":"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586"} Dec 10 14:49:12 crc kubenswrapper[4847]: I1210 14:49:12.330935 4847 generic.go:334] "Generic (PLEG): container finished" podID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerID="de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586" exitCode=0 Dec 10 14:49:12 crc kubenswrapper[4847]: I1210 14:49:12.331011 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerDied","Data":"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586"} Dec 10 14:49:13 crc kubenswrapper[4847]: I1210 14:49:13.358380 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerStarted","Data":"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69"} Dec 10 14:49:13 crc kubenswrapper[4847]: I1210 14:49:13.386314 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-trb8d" podStartSLOduration=2.8462918950000002 podStartE2EDuration="5.386282001s" podCreationTimestamp="2025-12-10 14:49:08 +0000 UTC" firstStartedPulling="2025-12-10 14:49:10.305230213 +0000 UTC m=+1499.874447843" lastFinishedPulling="2025-12-10 14:49:12.845220319 +0000 UTC m=+1502.414437949" observedRunningTime="2025-12-10 14:49:13.378259083 +0000 UTC m=+1502.947476743" watchObservedRunningTime="2025-12-10 14:49:13.386282001 +0000 UTC m=+1502.955499621" Dec 10 14:49:19 crc kubenswrapper[4847]: I1210 14:49:19.049171 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:19 crc kubenswrapper[4847]: I1210 14:49:19.050830 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:19 crc kubenswrapper[4847]: I1210 14:49:19.114004 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:19 crc kubenswrapper[4847]: I1210 14:49:19.467121 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:19 crc kubenswrapper[4847]: I1210 14:49:19.523476 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.440349 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-trb8d" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="registry-server" containerID="cri-o://f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69" gracePeriod=2 Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.712885 4847 scope.go:117] "RemoveContainer" containerID="ef58603d18d1560101302959777cb2ad21ef5d3fe8abd79001769aa94efb0d2b" Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.797489 4847 scope.go:117] "RemoveContainer" containerID="7399057319f9de26ea7513b978760aa2f589a8804d8990d90a7990641b0fa320" Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.848877 4847 scope.go:117] "RemoveContainer" containerID="527171fd46b18e40948541c76cef5c6c38d973701d9bab2dcfd122cb3c8efa31" Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.933056 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.994476 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5gm7\" (UniqueName: \"kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7\") pod \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.994587 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities\") pod \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.994778 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content\") pod \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\" (UID: \"a4df3602-2c3c-4428-99e7-a40ab62cffe6\") " Dec 10 14:49:21 crc kubenswrapper[4847]: I1210 14:49:21.995772 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities" (OuterVolumeSpecName: "utilities") pod "a4df3602-2c3c-4428-99e7-a40ab62cffe6" (UID: "a4df3602-2c3c-4428-99e7-a40ab62cffe6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.002882 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7" (OuterVolumeSpecName: "kube-api-access-t5gm7") pod "a4df3602-2c3c-4428-99e7-a40ab62cffe6" (UID: "a4df3602-2c3c-4428-99e7-a40ab62cffe6"). InnerVolumeSpecName "kube-api-access-t5gm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.016611 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4df3602-2c3c-4428-99e7-a40ab62cffe6" (UID: "a4df3602-2c3c-4428-99e7-a40ab62cffe6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.098006 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.098056 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5gm7\" (UniqueName: \"kubernetes.io/projected/a4df3602-2c3c-4428-99e7-a40ab62cffe6-kube-api-access-t5gm7\") on node \"crc\" DevicePath \"\"" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.098073 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4df3602-2c3c-4428-99e7-a40ab62cffe6-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.457316 4847 generic.go:334] "Generic (PLEG): container finished" podID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerID="f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69" exitCode=0 Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.457385 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trb8d" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.457385 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerDied","Data":"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69"} Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.457480 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trb8d" event={"ID":"a4df3602-2c3c-4428-99e7-a40ab62cffe6","Type":"ContainerDied","Data":"543167e5dd769eda57b5df35ab3626abd99ae6fcde0daf60d7741f820674595b"} Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.457528 4847 scope.go:117] "RemoveContainer" containerID="f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.481174 4847 scope.go:117] "RemoveContainer" containerID="de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.499187 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.506865 4847 scope.go:117] "RemoveContainer" containerID="3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.508418 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-trb8d"] Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.531089 4847 scope.go:117] "RemoveContainer" containerID="f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69" Dec 10 14:49:22 crc kubenswrapper[4847]: E1210 14:49:22.531664 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69\": container with ID starting with f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69 not found: ID does not exist" containerID="f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.531708 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69"} err="failed to get container status \"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69\": rpc error: code = NotFound desc = could not find container \"f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69\": container with ID starting with f27c2710095cf3d1cfedfa8795140a9d1b7aaa5ed05f2e8ee540e23a83bd2b69 not found: ID does not exist" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.531757 4847 scope.go:117] "RemoveContainer" containerID="de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586" Dec 10 14:49:22 crc kubenswrapper[4847]: E1210 14:49:22.532339 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586\": container with ID starting with de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586 not found: ID does not exist" containerID="de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.532361 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586"} err="failed to get container status \"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586\": rpc error: code = NotFound desc = could not find container \"de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586\": container with ID starting with de32a6d7e5fe4d1e8db58c61a437a1036dd62605ff3f0a05704211dcd7249586 not found: ID does not exist" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.532376 4847 scope.go:117] "RemoveContainer" containerID="3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094" Dec 10 14:49:22 crc kubenswrapper[4847]: E1210 14:49:22.532832 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094\": container with ID starting with 3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094 not found: ID does not exist" containerID="3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.532898 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094"} err="failed to get container status \"3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094\": rpc error: code = NotFound desc = could not find container \"3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094\": container with ID starting with 3be5c5c00bd234f693ea1c252598e74ca2d46cf45bcedbceff1aecdf3ce82094 not found: ID does not exist" Dec 10 14:49:22 crc kubenswrapper[4847]: I1210 14:49:22.779044 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" path="/var/lib/kubelet/pods/a4df3602-2c3c-4428-99e7-a40ab62cffe6/volumes" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.011021 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.011763 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.011847 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.013368 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.013452 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" gracePeriod=600 Dec 10 14:49:31 crc kubenswrapper[4847]: E1210 14:49:31.154792 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.547653 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" exitCode=0 Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.547702 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e"} Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.547750 4847 scope.go:117] "RemoveContainer" containerID="0a4cc06d642e3af5068983f2019bcb6d2829b2e1f89b4d5816a9fe9419ff4187" Dec 10 14:49:31 crc kubenswrapper[4847]: I1210 14:49:31.548634 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:49:31 crc kubenswrapper[4847]: E1210 14:49:31.548962 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:49:46 crc kubenswrapper[4847]: I1210 14:49:46.760870 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:49:46 crc kubenswrapper[4847]: E1210 14:49:46.762193 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.527297 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s7zpg"] Dec 10 14:49:53 crc kubenswrapper[4847]: E1210 14:49:53.529452 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="registry-server" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.529654 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="registry-server" Dec 10 14:49:53 crc kubenswrapper[4847]: E1210 14:49:53.529692 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="extract-content" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.529703 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="extract-content" Dec 10 14:49:53 crc kubenswrapper[4847]: E1210 14:49:53.529744 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="extract-utilities" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.529752 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="extract-utilities" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.530048 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4df3602-2c3c-4428-99e7-a40ab62cffe6" containerName="registry-server" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.532097 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.544436 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7zpg"] Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.596852 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwfhw\" (UniqueName: \"kubernetes.io/projected/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-kube-api-access-pwfhw\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.597313 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-catalog-content\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.597516 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-utilities\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.699400 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-utilities\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.699497 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwfhw\" (UniqueName: \"kubernetes.io/projected/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-kube-api-access-pwfhw\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.699638 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-catalog-content\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.700046 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-utilities\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.700066 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-catalog-content\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.719915 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwfhw\" (UniqueName: \"kubernetes.io/projected/ad6ec978-5852-4454-a0ef-b4d75cbe12d2-kube-api-access-pwfhw\") pod \"community-operators-s7zpg\" (UID: \"ad6ec978-5852-4454-a0ef-b4d75cbe12d2\") " pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:53 crc kubenswrapper[4847]: I1210 14:49:53.860669 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:49:54 crc kubenswrapper[4847]: I1210 14:49:54.516568 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7zpg"] Dec 10 14:49:54 crc kubenswrapper[4847]: I1210 14:49:54.782214 4847 generic.go:334] "Generic (PLEG): container finished" podID="ad6ec978-5852-4454-a0ef-b4d75cbe12d2" containerID="6f74504d9e2fcb45f102ddbec913ec7ccbd10363351cc55925bc56328683cfd2" exitCode=0 Dec 10 14:49:54 crc kubenswrapper[4847]: I1210 14:49:54.782268 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7zpg" event={"ID":"ad6ec978-5852-4454-a0ef-b4d75cbe12d2","Type":"ContainerDied","Data":"6f74504d9e2fcb45f102ddbec913ec7ccbd10363351cc55925bc56328683cfd2"} Dec 10 14:49:54 crc kubenswrapper[4847]: I1210 14:49:54.782316 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7zpg" event={"ID":"ad6ec978-5852-4454-a0ef-b4d75cbe12d2","Type":"ContainerStarted","Data":"c7352f0baf84e9cd2beee40631f780c38bff09696d8860cd9ce0e920413c9666"} Dec 10 14:49:58 crc kubenswrapper[4847]: I1210 14:49:58.760812 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:49:58 crc kubenswrapper[4847]: E1210 14:49:58.762048 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:49:59 crc kubenswrapper[4847]: I1210 14:49:59.833457 4847 generic.go:334] "Generic (PLEG): container finished" podID="ad6ec978-5852-4454-a0ef-b4d75cbe12d2" containerID="342832e2b9f416c6d3e7b8d11ea1d3c73e6d9a032fc38bf80a016fd9b6459d08" exitCode=0 Dec 10 14:49:59 crc kubenswrapper[4847]: I1210 14:49:59.833775 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7zpg" event={"ID":"ad6ec978-5852-4454-a0ef-b4d75cbe12d2","Type":"ContainerDied","Data":"342832e2b9f416c6d3e7b8d11ea1d3c73e6d9a032fc38bf80a016fd9b6459d08"} Dec 10 14:50:02 crc kubenswrapper[4847]: I1210 14:50:02.863778 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7zpg" event={"ID":"ad6ec978-5852-4454-a0ef-b4d75cbe12d2","Type":"ContainerStarted","Data":"870ddca8eb0e0e3c8ba0af2a3c1bb6cf40a6286725d2d490e300b93c227d5b22"} Dec 10 14:50:02 crc kubenswrapper[4847]: I1210 14:50:02.896290 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s7zpg" podStartSLOduration=2.149111519 podStartE2EDuration="9.896267584s" podCreationTimestamp="2025-12-10 14:49:53 +0000 UTC" firstStartedPulling="2025-12-10 14:49:54.784146575 +0000 UTC m=+1544.353364215" lastFinishedPulling="2025-12-10 14:50:02.53130265 +0000 UTC m=+1552.100520280" observedRunningTime="2025-12-10 14:50:02.889952982 +0000 UTC m=+1552.459170632" watchObservedRunningTime="2025-12-10 14:50:02.896267584 +0000 UTC m=+1552.465485214" Dec 10 14:50:03 crc kubenswrapper[4847]: I1210 14:50:03.861653 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:50:03 crc kubenswrapper[4847]: I1210 14:50:03.861691 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:50:04 crc kubenswrapper[4847]: I1210 14:50:04.920669 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s7zpg" podUID="ad6ec978-5852-4454-a0ef-b4d75cbe12d2" containerName="registry-server" probeResult="failure" output=< Dec 10 14:50:04 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:50:04 crc kubenswrapper[4847]: > Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.543019 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.547059 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.554166 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.653824 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.654253 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvvz\" (UniqueName: \"kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.654435 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.756826 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.756921 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvvz\" (UniqueName: \"kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.757001 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.757359 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.757528 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.776598 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvvz\" (UniqueName: \"kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz\") pod \"certified-operators-qz856\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:10 crc kubenswrapper[4847]: I1210 14:50:10.869609 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:11 crc kubenswrapper[4847]: I1210 14:50:11.504187 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:11 crc kubenswrapper[4847]: I1210 14:50:11.759510 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:50:11 crc kubenswrapper[4847]: E1210 14:50:11.760332 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:50:11 crc kubenswrapper[4847]: I1210 14:50:11.959643 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerStarted","Data":"ec0f489b5aa3a9da10c35d538c28a7de628beb9c356a5b24d7556a169fa23245"} Dec 10 14:50:12 crc kubenswrapper[4847]: I1210 14:50:12.968387 4847 generic.go:334] "Generic (PLEG): container finished" podID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerID="d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3" exitCode=0 Dec 10 14:50:12 crc kubenswrapper[4847]: I1210 14:50:12.968467 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerDied","Data":"d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3"} Dec 10 14:50:13 crc kubenswrapper[4847]: I1210 14:50:13.915190 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:50:13 crc kubenswrapper[4847]: I1210 14:50:13.982614 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerStarted","Data":"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b"} Dec 10 14:50:13 crc kubenswrapper[4847]: I1210 14:50:13.984295 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s7zpg" Dec 10 14:50:14 crc kubenswrapper[4847]: I1210 14:50:14.992726 4847 generic.go:334] "Generic (PLEG): container finished" podID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerID="250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b" exitCode=0 Dec 10 14:50:14 crc kubenswrapper[4847]: I1210 14:50:14.992769 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerDied","Data":"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b"} Dec 10 14:50:15 crc kubenswrapper[4847]: I1210 14:50:15.338460 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7zpg"] Dec 10 14:50:15 crc kubenswrapper[4847]: I1210 14:50:15.732503 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:50:15 crc kubenswrapper[4847]: I1210 14:50:15.732983 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r78kt" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="registry-server" containerID="cri-o://af10217921422cc7ea45c3ab722300a5a8278e5a3fe138f2828840eb7ec03804" gracePeriod=2 Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.023235 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerStarted","Data":"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0"} Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.038090 4847 generic.go:334] "Generic (PLEG): container finished" podID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerID="af10217921422cc7ea45c3ab722300a5a8278e5a3fe138f2828840eb7ec03804" exitCode=0 Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.038129 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerDied","Data":"af10217921422cc7ea45c3ab722300a5a8278e5a3fe138f2828840eb7ec03804"} Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.055347 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qz856" podStartSLOduration=3.515495032 podStartE2EDuration="6.055327815s" podCreationTimestamp="2025-12-10 14:50:10 +0000 UTC" firstStartedPulling="2025-12-10 14:50:12.970524254 +0000 UTC m=+1562.539741884" lastFinishedPulling="2025-12-10 14:50:15.510357037 +0000 UTC m=+1565.079574667" observedRunningTime="2025-12-10 14:50:16.047992816 +0000 UTC m=+1565.617210446" watchObservedRunningTime="2025-12-10 14:50:16.055327815 +0000 UTC m=+1565.624545445" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.298747 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.378230 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content\") pod \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.378302 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities\") pod \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.378428 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npgts\" (UniqueName: \"kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts\") pod \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\" (UID: \"68557631-c526-4bcf-8f5a-0f5b9e89a3f9\") " Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.379052 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities" (OuterVolumeSpecName: "utilities") pod "68557631-c526-4bcf-8f5a-0f5b9e89a3f9" (UID: "68557631-c526-4bcf-8f5a-0f5b9e89a3f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.390978 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts" (OuterVolumeSpecName: "kube-api-access-npgts") pod "68557631-c526-4bcf-8f5a-0f5b9e89a3f9" (UID: "68557631-c526-4bcf-8f5a-0f5b9e89a3f9"). InnerVolumeSpecName "kube-api-access-npgts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.436756 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68557631-c526-4bcf-8f5a-0f5b9e89a3f9" (UID: "68557631-c526-4bcf-8f5a-0f5b9e89a3f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.480449 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.480502 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:16 crc kubenswrapper[4847]: I1210 14:50:16.480518 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npgts\" (UniqueName: \"kubernetes.io/projected/68557631-c526-4bcf-8f5a-0f5b9e89a3f9-kube-api-access-npgts\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.068893 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r78kt" event={"ID":"68557631-c526-4bcf-8f5a-0f5b9e89a3f9","Type":"ContainerDied","Data":"4491b236d50cf26108cea387cd75904aa9ddd2b60afe590925c6e245521f35f8"} Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.068978 4847 scope.go:117] "RemoveContainer" containerID="af10217921422cc7ea45c3ab722300a5a8278e5a3fe138f2828840eb7ec03804" Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.068972 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r78kt" Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.102997 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.120433 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r78kt"] Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.121470 4847 scope.go:117] "RemoveContainer" containerID="7cf7549138c29660d9cf9f62372db2f96746ff66efa71003242f0f6559f4e2b8" Dec 10 14:50:17 crc kubenswrapper[4847]: I1210 14:50:17.141603 4847 scope.go:117] "RemoveContainer" containerID="9ad7113fb685bd443625d6a3a2705366f5539129a48900acd4eb6189731f0fef" Dec 10 14:50:18 crc kubenswrapper[4847]: I1210 14:50:18.770856 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" path="/var/lib/kubelet/pods/68557631-c526-4bcf-8f5a-0f5b9e89a3f9/volumes" Dec 10 14:50:20 crc kubenswrapper[4847]: I1210 14:50:20.870441 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:20 crc kubenswrapper[4847]: I1210 14:50:20.870853 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:20 crc kubenswrapper[4847]: I1210 14:50:20.969121 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:21 crc kubenswrapper[4847]: I1210 14:50:21.154305 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:22 crc kubenswrapper[4847]: I1210 14:50:22.022534 4847 scope.go:117] "RemoveContainer" containerID="d3c0fad1dee22d2a9d92e55d0f6fdc48d84ff1c2f9fa712011a28308fed14a33" Dec 10 14:50:22 crc kubenswrapper[4847]: I1210 14:50:22.046821 4847 scope.go:117] "RemoveContainer" containerID="a710b931fedbf0a85ddad2347e83365594a0b23b9583983cc7f44422aaaac81e" Dec 10 14:50:22 crc kubenswrapper[4847]: I1210 14:50:22.125366 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:22 crc kubenswrapper[4847]: I1210 14:50:22.760186 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:50:22 crc kubenswrapper[4847]: E1210 14:50:22.760914 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.128630 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qz856" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="registry-server" containerID="cri-o://aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0" gracePeriod=2 Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.600794 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.717073 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities\") pod \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.717148 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content\") pod \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.717225 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvvvz\" (UniqueName: \"kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz\") pod \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\" (UID: \"ccf802b6-a0ba-4e19-aed4-b13741c4321a\") " Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.718397 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities" (OuterVolumeSpecName: "utilities") pod "ccf802b6-a0ba-4e19-aed4-b13741c4321a" (UID: "ccf802b6-a0ba-4e19-aed4-b13741c4321a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.726893 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz" (OuterVolumeSpecName: "kube-api-access-lvvvz") pod "ccf802b6-a0ba-4e19-aed4-b13741c4321a" (UID: "ccf802b6-a0ba-4e19-aed4-b13741c4321a"). InnerVolumeSpecName "kube-api-access-lvvvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.775394 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccf802b6-a0ba-4e19-aed4-b13741c4321a" (UID: "ccf802b6-a0ba-4e19-aed4-b13741c4321a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.819500 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.819524 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccf802b6-a0ba-4e19-aed4-b13741c4321a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:23 crc kubenswrapper[4847]: I1210 14:50:23.819537 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvvvz\" (UniqueName: \"kubernetes.io/projected/ccf802b6-a0ba-4e19-aed4-b13741c4321a-kube-api-access-lvvvz\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.139327 4847 generic.go:334] "Generic (PLEG): container finished" podID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerID="aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0" exitCode=0 Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.139380 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerDied","Data":"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0"} Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.139404 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qz856" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.139420 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qz856" event={"ID":"ccf802b6-a0ba-4e19-aed4-b13741c4321a","Type":"ContainerDied","Data":"ec0f489b5aa3a9da10c35d538c28a7de628beb9c356a5b24d7556a169fa23245"} Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.139447 4847 scope.go:117] "RemoveContainer" containerID="aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.160577 4847 scope.go:117] "RemoveContainer" containerID="250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.172590 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.181004 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qz856"] Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.190966 4847 scope.go:117] "RemoveContainer" containerID="d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.229909 4847 scope.go:117] "RemoveContainer" containerID="aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0" Dec 10 14:50:24 crc kubenswrapper[4847]: E1210 14:50:24.230311 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0\": container with ID starting with aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0 not found: ID does not exist" containerID="aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.230347 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0"} err="failed to get container status \"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0\": rpc error: code = NotFound desc = could not find container \"aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0\": container with ID starting with aa1973bde240c1f15e077594af6ff3d54848bed6af0250783c609cbeb28fc5c0 not found: ID does not exist" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.230374 4847 scope.go:117] "RemoveContainer" containerID="250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b" Dec 10 14:50:24 crc kubenswrapper[4847]: E1210 14:50:24.230906 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b\": container with ID starting with 250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b not found: ID does not exist" containerID="250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.230928 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b"} err="failed to get container status \"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b\": rpc error: code = NotFound desc = could not find container \"250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b\": container with ID starting with 250a3ef8e41e0ee1838491c9b6797615de071bbbf91d917c6a0e781c334dac6b not found: ID does not exist" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.230943 4847 scope.go:117] "RemoveContainer" containerID="d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3" Dec 10 14:50:24 crc kubenswrapper[4847]: E1210 14:50:24.231234 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3\": container with ID starting with d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3 not found: ID does not exist" containerID="d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.231260 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3"} err="failed to get container status \"d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3\": rpc error: code = NotFound desc = could not find container \"d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3\": container with ID starting with d79eb5388f3b4479a3f7d4e6feee2fc31a3df6d379062457b6917eddabfb85f3 not found: ID does not exist" Dec 10 14:50:24 crc kubenswrapper[4847]: I1210 14:50:24.770316 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" path="/var/lib/kubelet/pods/ccf802b6-a0ba-4e19-aed4-b13741c4321a/volumes" Dec 10 14:50:37 crc kubenswrapper[4847]: I1210 14:50:37.760408 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:50:37 crc kubenswrapper[4847]: E1210 14:50:37.762425 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:50:41 crc kubenswrapper[4847]: I1210 14:50:41.302702 4847 generic.go:334] "Generic (PLEG): container finished" podID="c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" containerID="7ef3cf3ed8df4801c6249a157fc453aa18140c35aaa37903117854e9f6f70972" exitCode=0 Dec 10 14:50:41 crc kubenswrapper[4847]: I1210 14:50:41.302884 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" event={"ID":"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6","Type":"ContainerDied","Data":"7ef3cf3ed8df4801c6249a157fc453aa18140c35aaa37903117854e9f6f70972"} Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.698110 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.777092 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key\") pod \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.777481 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle\") pod \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.777606 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory\") pod \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.777644 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gffl4\" (UniqueName: \"kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4\") pod \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\" (UID: \"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6\") " Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.789917 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" (UID: "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.809038 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4" (OuterVolumeSpecName: "kube-api-access-gffl4") pod "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" (UID: "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6"). InnerVolumeSpecName "kube-api-access-gffl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.819996 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" (UID: "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.879538 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gffl4\" (UniqueName: \"kubernetes.io/projected/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-kube-api-access-gffl4\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.879573 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.879584 4847 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.893885 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory" (OuterVolumeSpecName: "inventory") pod "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" (UID: "c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:50:42 crc kubenswrapper[4847]: I1210 14:50:42.980896 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.324226 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" event={"ID":"c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6","Type":"ContainerDied","Data":"729e0e896ea151ca6317345285f24f1876f81ae2f21083224decda0b7b3d2cfa"} Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.324277 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="729e0e896ea151ca6317345285f24f1876f81ae2f21083224decda0b7b3d2cfa" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.324279 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411093 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h"] Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411660 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="extract-utilities" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411694 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="extract-utilities" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411735 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="extract-content" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411745 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="extract-content" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411765 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411774 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411811 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="extract-utilities" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411828 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="extract-utilities" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411857 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411886 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411906 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="extract-content" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411918 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="extract-content" Dec 10 14:50:43 crc kubenswrapper[4847]: E1210 14:50:43.411937 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.411950 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.412205 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf802b6-a0ba-4e19-aed4-b13741c4321a" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.412246 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.412266 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="68557631-c526-4bcf-8f5a-0f5b9e89a3f9" containerName="registry-server" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.413122 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.416653 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.417066 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.417392 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.420355 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.421649 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h"] Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.491100 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.491458 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.491498 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzdpb\" (UniqueName: \"kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.593419 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.593480 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.593516 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzdpb\" (UniqueName: \"kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.598874 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.599106 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.611442 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzdpb\" (UniqueName: \"kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-krj6h\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:43 crc kubenswrapper[4847]: I1210 14:50:43.734920 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:50:44 crc kubenswrapper[4847]: I1210 14:50:44.246739 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h"] Dec 10 14:50:44 crc kubenswrapper[4847]: I1210 14:50:44.335915 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" event={"ID":"48c5aa38-4902-4105-99ac-40fb8ab5a288","Type":"ContainerStarted","Data":"366c28facdd6adba41c6f4784539c17ec9741413d72a385b6f7969bc98552760"} Dec 10 14:50:45 crc kubenswrapper[4847]: I1210 14:50:45.369378 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" event={"ID":"48c5aa38-4902-4105-99ac-40fb8ab5a288","Type":"ContainerStarted","Data":"6a1925092987c97a5a721808284dab6c36a551e844ded7904994934930e7ff2a"} Dec 10 14:50:45 crc kubenswrapper[4847]: I1210 14:50:45.399332 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" podStartSLOduration=1.8891425320000002 podStartE2EDuration="2.399312592s" podCreationTimestamp="2025-12-10 14:50:43 +0000 UTC" firstStartedPulling="2025-12-10 14:50:44.254510581 +0000 UTC m=+1593.823728211" lastFinishedPulling="2025-12-10 14:50:44.764680641 +0000 UTC m=+1594.333898271" observedRunningTime="2025-12-10 14:50:45.387767288 +0000 UTC m=+1594.956984938" watchObservedRunningTime="2025-12-10 14:50:45.399312592 +0000 UTC m=+1594.968530222" Dec 10 14:50:48 crc kubenswrapper[4847]: I1210 14:50:48.760283 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:50:48 crc kubenswrapper[4847]: E1210 14:50:48.761328 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:01 crc kubenswrapper[4847]: I1210 14:51:01.759375 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:51:01 crc kubenswrapper[4847]: E1210 14:51:01.760275 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:15 crc kubenswrapper[4847]: I1210 14:51:15.759449 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:51:15 crc kubenswrapper[4847]: E1210 14:51:15.760285 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:29 crc kubenswrapper[4847]: I1210 14:51:29.760113 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:51:29 crc kubenswrapper[4847]: E1210 14:51:29.760833 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:41 crc kubenswrapper[4847]: I1210 14:51:41.760878 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:51:41 crc kubenswrapper[4847]: E1210 14:51:41.761769 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:52 crc kubenswrapper[4847]: I1210 14:51:52.759904 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:51:52 crc kubenswrapper[4847]: E1210 14:51:52.760729 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:51:59 crc kubenswrapper[4847]: I1210 14:51:59.041640 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2l5wx"] Dec 10 14:51:59 crc kubenswrapper[4847]: I1210 14:51:59.050620 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2595-account-create-update-hxnfg"] Dec 10 14:51:59 crc kubenswrapper[4847]: I1210 14:51:59.060256 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2l5wx"] Dec 10 14:51:59 crc kubenswrapper[4847]: I1210 14:51:59.069509 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2595-account-create-update-hxnfg"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.036576 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a9e3-account-create-update-8s2rf"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.046386 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jgdv2"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.058447 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-sj7rt"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.071265 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-260c-account-create-update-cnfqw"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.083743 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-sj7rt"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.095604 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a9e3-account-create-update-8s2rf"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.107234 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-260c-account-create-update-cnfqw"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.117362 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jgdv2"] Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.772821 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06fc695e-403f-4858-adfc-0ed1f8d414c7" path="/var/lib/kubelet/pods/06fc695e-403f-4858-adfc-0ed1f8d414c7/volumes" Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.773489 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e3cce4-5609-460e-b5b1-8e8a17962bc9" path="/var/lib/kubelet/pods/17e3cce4-5609-460e-b5b1-8e8a17962bc9/volumes" Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.774204 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8711478e-be68-44d2-9471-225c0008b966" path="/var/lib/kubelet/pods/8711478e-be68-44d2-9471-225c0008b966/volumes" Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.774819 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b0fdd70-30cd-49a4-8072-044ef1cfd27a" path="/var/lib/kubelet/pods/8b0fdd70-30cd-49a4-8072-044ef1cfd27a/volumes" Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.775887 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="995e809e-0a8d-4b1e-acd8-b64571571651" path="/var/lib/kubelet/pods/995e809e-0a8d-4b1e-acd8-b64571571651/volumes" Dec 10 14:52:00 crc kubenswrapper[4847]: I1210 14:52:00.776446 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abac02be-a75e-4ba2-ac00-4cb338681fec" path="/var/lib/kubelet/pods/abac02be-a75e-4ba2-ac00-4cb338681fec/volumes" Dec 10 14:52:06 crc kubenswrapper[4847]: I1210 14:52:06.759983 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:52:06 crc kubenswrapper[4847]: E1210 14:52:06.760801 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:52:18 crc kubenswrapper[4847]: I1210 14:52:18.760312 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:52:18 crc kubenswrapper[4847]: E1210 14:52:18.761097 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:52:20 crc kubenswrapper[4847]: I1210 14:52:20.293244 4847 generic.go:334] "Generic (PLEG): container finished" podID="48c5aa38-4902-4105-99ac-40fb8ab5a288" containerID="6a1925092987c97a5a721808284dab6c36a551e844ded7904994934930e7ff2a" exitCode=0 Dec 10 14:52:20 crc kubenswrapper[4847]: I1210 14:52:20.293359 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" event={"ID":"48c5aa38-4902-4105-99ac-40fb8ab5a288","Type":"ContainerDied","Data":"6a1925092987c97a5a721808284dab6c36a551e844ded7904994934930e7ff2a"} Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.697007 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.816587 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzdpb\" (UniqueName: \"kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb\") pod \"48c5aa38-4902-4105-99ac-40fb8ab5a288\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.816760 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key\") pod \"48c5aa38-4902-4105-99ac-40fb8ab5a288\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.816820 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory\") pod \"48c5aa38-4902-4105-99ac-40fb8ab5a288\" (UID: \"48c5aa38-4902-4105-99ac-40fb8ab5a288\") " Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.830848 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb" (OuterVolumeSpecName: "kube-api-access-vzdpb") pod "48c5aa38-4902-4105-99ac-40fb8ab5a288" (UID: "48c5aa38-4902-4105-99ac-40fb8ab5a288"). InnerVolumeSpecName "kube-api-access-vzdpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.858298 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48c5aa38-4902-4105-99ac-40fb8ab5a288" (UID: "48c5aa38-4902-4105-99ac-40fb8ab5a288"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.870994 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory" (OuterVolumeSpecName: "inventory") pod "48c5aa38-4902-4105-99ac-40fb8ab5a288" (UID: "48c5aa38-4902-4105-99ac-40fb8ab5a288"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.923410 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.923448 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c5aa38-4902-4105-99ac-40fb8ab5a288-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:52:21 crc kubenswrapper[4847]: I1210 14:52:21.923458 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzdpb\" (UniqueName: \"kubernetes.io/projected/48c5aa38-4902-4105-99ac-40fb8ab5a288-kube-api-access-vzdpb\") on node \"crc\" DevicePath \"\"" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.032078 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-cc2e-account-create-update-qz9b9"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.047168 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-906f-account-create-update-xk8nx"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.056703 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-fbd72"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.064732 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-906f-account-create-update-xk8nx"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.072770 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-cc2e-account-create-update-qz9b9"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.080861 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-fbd72"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.209672 4847 scope.go:117] "RemoveContainer" containerID="2fa46b3bcdfe3f7db9c04e77b2a3d8fb6d1a72530ce7e1a9e77b3ff1348e3d66" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.234260 4847 scope.go:117] "RemoveContainer" containerID="365bb82dbca333a1c0be98734a242477ddd0290bb75bb639cfbc566d1dd8b5d5" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.274689 4847 scope.go:117] "RemoveContainer" containerID="dab5b9227e9761db9e739272db5d831ff23121f2dafb0cd2ff29f353e3e20b4c" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.292917 4847 scope.go:117] "RemoveContainer" containerID="a3da6562e66727d6f12a22f77c2c1e7013586e298df7b61811b075d98ad79966" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.318167 4847 scope.go:117] "RemoveContainer" containerID="358cdbf59419eac3710fd1b5c8c00a928ab6b639e9132906ac3d51c6355ddfd9" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.319180 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" event={"ID":"48c5aa38-4902-4105-99ac-40fb8ab5a288","Type":"ContainerDied","Data":"366c28facdd6adba41c6f4784539c17ec9741413d72a385b6f7969bc98552760"} Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.319290 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="366c28facdd6adba41c6f4784539c17ec9741413d72a385b6f7969bc98552760" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.319223 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-krj6h" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.365120 4847 scope.go:117] "RemoveContainer" containerID="b06a4da59be924173fb8a1d5b7e8c78a26a0174a4fd641e2a61605c216ee093e" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.410881 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf"] Dec 10 14:52:22 crc kubenswrapper[4847]: E1210 14:52:22.411392 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c5aa38-4902-4105-99ac-40fb8ab5a288" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.411415 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c5aa38-4902-4105-99ac-40fb8ab5a288" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.411652 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c5aa38-4902-4105-99ac-40fb8ab5a288" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.412444 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.415223 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.415460 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.415609 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.421420 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.445875 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf"] Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.536056 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fxmf\" (UniqueName: \"kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.536615 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.536756 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.638800 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fxmf\" (UniqueName: \"kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.638895 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.639043 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.645075 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.652239 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.655033 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fxmf\" (UniqueName: \"kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.741284 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.774814 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17bbab02-b41c-4b58-a235-0c88b7eeb1b3" path="/var/lib/kubelet/pods/17bbab02-b41c-4b58-a235-0c88b7eeb1b3/volumes" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.779439 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a11592b8-0374-4b1d-a579-64b72830148a" path="/var/lib/kubelet/pods/a11592b8-0374-4b1d-a579-64b72830148a/volumes" Dec 10 14:52:22 crc kubenswrapper[4847]: I1210 14:52:22.780180 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5096792-9e2e-4843-85da-9e14112d056a" path="/var/lib/kubelet/pods/e5096792-9e2e-4843-85da-9e14112d056a/volumes" Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.039914 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-4cxxl"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.052184 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bd7a-account-create-update-l2vzd"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.059879 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bb6jt"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.066963 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bb6jt"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.074845 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-4cxxl"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.082859 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bd7a-account-create-update-l2vzd"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.291095 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf"] Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.301516 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:52:23 crc kubenswrapper[4847]: I1210 14:52:23.334157 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" event={"ID":"21486092-46dc-4bf8-9cfc-5fc5781461c2","Type":"ContainerStarted","Data":"c6fbf5d8ce4b9872a454a4903043022f96f123ce6f2fc598a301e240d9dc6212"} Dec 10 14:52:24 crc kubenswrapper[4847]: I1210 14:52:24.769812 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bba7667-cbb4-4636-9197-2409e86b8cde" path="/var/lib/kubelet/pods/0bba7667-cbb4-4636-9197-2409e86b8cde/volumes" Dec 10 14:52:24 crc kubenswrapper[4847]: I1210 14:52:24.770994 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b75aa68-b977-4614-877b-9828d3045226" path="/var/lib/kubelet/pods/3b75aa68-b977-4614-877b-9828d3045226/volumes" Dec 10 14:52:24 crc kubenswrapper[4847]: I1210 14:52:24.771558 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7148ddb7-2803-41ab-bf95-f820c60614d1" path="/var/lib/kubelet/pods/7148ddb7-2803-41ab-bf95-f820c60614d1/volumes" Dec 10 14:52:25 crc kubenswrapper[4847]: I1210 14:52:25.362280 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" event={"ID":"21486092-46dc-4bf8-9cfc-5fc5781461c2","Type":"ContainerStarted","Data":"9ce37199d4d528005a9b10738101397c20fe4e9c1035e92aaad67cfb1285e62e"} Dec 10 14:52:31 crc kubenswrapper[4847]: I1210 14:52:31.759007 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:52:31 crc kubenswrapper[4847]: E1210 14:52:31.759807 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:52:37 crc kubenswrapper[4847]: I1210 14:52:37.039841 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" podStartSLOduration=13.76298313 podStartE2EDuration="15.039818861s" podCreationTimestamp="2025-12-10 14:52:22 +0000 UTC" firstStartedPulling="2025-12-10 14:52:23.301203809 +0000 UTC m=+1692.870421439" lastFinishedPulling="2025-12-10 14:52:24.57803952 +0000 UTC m=+1694.147257170" observedRunningTime="2025-12-10 14:52:25.387958425 +0000 UTC m=+1694.957176075" watchObservedRunningTime="2025-12-10 14:52:37.039818861 +0000 UTC m=+1706.609036501" Dec 10 14:52:37 crc kubenswrapper[4847]: I1210 14:52:37.041258 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ssqhj"] Dec 10 14:52:37 crc kubenswrapper[4847]: I1210 14:52:37.051671 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ssqhj"] Dec 10 14:52:38 crc kubenswrapper[4847]: I1210 14:52:38.774837 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecf994d1-33cd-404a-bec2-01956b20b78d" path="/var/lib/kubelet/pods/ecf994d1-33cd-404a-bec2-01956b20b78d/volumes" Dec 10 14:52:45 crc kubenswrapper[4847]: I1210 14:52:45.760387 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:52:45 crc kubenswrapper[4847]: E1210 14:52:45.761009 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:52:57 crc kubenswrapper[4847]: I1210 14:52:57.761426 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:52:57 crc kubenswrapper[4847]: E1210 14:52:57.762820 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:53:10 crc kubenswrapper[4847]: I1210 14:53:10.841431 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:53:10 crc kubenswrapper[4847]: E1210 14:53:10.882924 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:53:11 crc kubenswrapper[4847]: I1210 14:53:11.060876 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-czhlz"] Dec 10 14:53:11 crc kubenswrapper[4847]: I1210 14:53:11.070204 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-czhlz"] Dec 10 14:53:12 crc kubenswrapper[4847]: I1210 14:53:12.025540 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-66gqz"] Dec 10 14:53:12 crc kubenswrapper[4847]: I1210 14:53:12.033670 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-66gqz"] Dec 10 14:53:12 crc kubenswrapper[4847]: I1210 14:53:12.772725 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32400504-9741-4abf-bdbf-a8ea9fcb8e4e" path="/var/lib/kubelet/pods/32400504-9741-4abf-bdbf-a8ea9fcb8e4e/volumes" Dec 10 14:53:12 crc kubenswrapper[4847]: I1210 14:53:12.773652 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4071c9-d190-4706-8f11-16b77f3db069" path="/var/lib/kubelet/pods/ea4071c9-d190-4706-8f11-16b77f3db069/volumes" Dec 10 14:53:21 crc kubenswrapper[4847]: I1210 14:53:21.759836 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:53:21 crc kubenswrapper[4847]: E1210 14:53:21.760876 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.540602 4847 scope.go:117] "RemoveContainer" containerID="cbbbc02925db69af15abd144ed9d350cb7ada6712454fc52a9a48c0cb9f476ad" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.615033 4847 scope.go:117] "RemoveContainer" containerID="84a54dec72b3da41bc66d1467dc78819f13c2500cbc38621a3e4180fd8e79163" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.649642 4847 scope.go:117] "RemoveContainer" containerID="80ae245596e55e80cde46d6afdd39ba3f0991e25a7959f972b9d4db4507ebafc" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.698001 4847 scope.go:117] "RemoveContainer" containerID="8d58579c90642bdd77f103403ec3a5e487b8e23736d073c9ee80a295d4f82afe" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.745950 4847 scope.go:117] "RemoveContainer" containerID="8347dbf81b888a31f98854a1ff7e74c2100b2619c2a2325b51d66e1270c79abf" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.831211 4847 scope.go:117] "RemoveContainer" containerID="c44e9fc63261d85023464a8913b4fbe3f982017c538e9dd12263d29e039c4a61" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.858385 4847 scope.go:117] "RemoveContainer" containerID="ecfca552d6bf41de43d26684aab0607b227fc027286c65be3ae47cce6d2b7a97" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.881669 4847 scope.go:117] "RemoveContainer" containerID="6ec3a10674aa81d3e1e8db5c312a2b5019e8ebd2983178875a1545c9454cdffa" Dec 10 14:53:22 crc kubenswrapper[4847]: I1210 14:53:22.937322 4847 scope.go:117] "RemoveContainer" containerID="99bd80ca3b3f97d20133979a88bec62c4a665c1b6ca7f05b1da592b91b6e6d6a" Dec 10 14:53:25 crc kubenswrapper[4847]: I1210 14:53:25.038762 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-j5hjt"] Dec 10 14:53:25 crc kubenswrapper[4847]: I1210 14:53:25.054351 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-j5hjt"] Dec 10 14:53:26 crc kubenswrapper[4847]: I1210 14:53:26.769943 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e3440ea-24b3-4cc9-9872-8c4e28ab340b" path="/var/lib/kubelet/pods/8e3440ea-24b3-4cc9-9872-8c4e28ab340b/volumes" Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.047328 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-p6r7l"] Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.065989 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fw2fp"] Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.082535 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-p6r7l"] Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.102725 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fw2fp"] Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.770083 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66a6b89e-e44e-4e19-92bd-12d58ad99fd1" path="/var/lib/kubelet/pods/66a6b89e-e44e-4e19-92bd-12d58ad99fd1/volumes" Dec 10 14:53:28 crc kubenswrapper[4847]: I1210 14:53:28.770665 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac62992-6083-4fa3-bed2-587f24f6bb40" path="/var/lib/kubelet/pods/aac62992-6083-4fa3-bed2-587f24f6bb40/volumes" Dec 10 14:53:36 crc kubenswrapper[4847]: I1210 14:53:36.759786 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:53:36 crc kubenswrapper[4847]: E1210 14:53:36.760556 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:53:39 crc kubenswrapper[4847]: I1210 14:53:39.055477 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-twlv7"] Dec 10 14:53:39 crc kubenswrapper[4847]: I1210 14:53:39.068128 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-twlv7"] Dec 10 14:53:40 crc kubenswrapper[4847]: I1210 14:53:40.773237 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ca7b48-ad83-46b4-a813-de1ec043a3e2" path="/var/lib/kubelet/pods/82ca7b48-ad83-46b4-a813-de1ec043a3e2/volumes" Dec 10 14:53:43 crc kubenswrapper[4847]: I1210 14:53:43.220128 4847 generic.go:334] "Generic (PLEG): container finished" podID="21486092-46dc-4bf8-9cfc-5fc5781461c2" containerID="9ce37199d4d528005a9b10738101397c20fe4e9c1035e92aaad67cfb1285e62e" exitCode=0 Dec 10 14:53:43 crc kubenswrapper[4847]: I1210 14:53:43.220211 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" event={"ID":"21486092-46dc-4bf8-9cfc-5fc5781461c2","Type":"ContainerDied","Data":"9ce37199d4d528005a9b10738101397c20fe4e9c1035e92aaad67cfb1285e62e"} Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.676209 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.766463 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key\") pod \"21486092-46dc-4bf8-9cfc-5fc5781461c2\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.767305 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fxmf\" (UniqueName: \"kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf\") pod \"21486092-46dc-4bf8-9cfc-5fc5781461c2\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.767492 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory\") pod \"21486092-46dc-4bf8-9cfc-5fc5781461c2\" (UID: \"21486092-46dc-4bf8-9cfc-5fc5781461c2\") " Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.776942 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf" (OuterVolumeSpecName: "kube-api-access-8fxmf") pod "21486092-46dc-4bf8-9cfc-5fc5781461c2" (UID: "21486092-46dc-4bf8-9cfc-5fc5781461c2"). InnerVolumeSpecName "kube-api-access-8fxmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.804979 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory" (OuterVolumeSpecName: "inventory") pod "21486092-46dc-4bf8-9cfc-5fc5781461c2" (UID: "21486092-46dc-4bf8-9cfc-5fc5781461c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.805546 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21486092-46dc-4bf8-9cfc-5fc5781461c2" (UID: "21486092-46dc-4bf8-9cfc-5fc5781461c2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.872054 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.872098 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fxmf\" (UniqueName: \"kubernetes.io/projected/21486092-46dc-4bf8-9cfc-5fc5781461c2-kube-api-access-8fxmf\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:44 crc kubenswrapper[4847]: I1210 14:53:44.872118 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21486092-46dc-4bf8-9cfc-5fc5781461c2-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.244372 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" event={"ID":"21486092-46dc-4bf8-9cfc-5fc5781461c2","Type":"ContainerDied","Data":"c6fbf5d8ce4b9872a454a4903043022f96f123ce6f2fc598a301e240d9dc6212"} Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.244917 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6fbf5d8ce4b9872a454a4903043022f96f123ce6f2fc598a301e240d9dc6212" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.244470 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.348393 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr"] Dec 10 14:53:45 crc kubenswrapper[4847]: E1210 14:53:45.349069 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21486092-46dc-4bf8-9cfc-5fc5781461c2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.349093 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="21486092-46dc-4bf8-9cfc-5fc5781461c2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.349424 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="21486092-46dc-4bf8-9cfc-5fc5781461c2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.350520 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.353650 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.353807 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.353812 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.358965 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.363432 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr"] Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.488501 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.488594 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2bz2\" (UniqueName: \"kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.489312 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.591107 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.591230 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.591280 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2bz2\" (UniqueName: \"kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.599658 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.601248 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.615884 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2bz2\" (UniqueName: \"kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:45 crc kubenswrapper[4847]: I1210 14:53:45.685473 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:46 crc kubenswrapper[4847]: I1210 14:53:46.272736 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr"] Dec 10 14:53:47 crc kubenswrapper[4847]: I1210 14:53:47.281588 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" event={"ID":"7a627a8d-ac7d-4d5d-9258-3342ca59930d","Type":"ContainerStarted","Data":"0276f0c6b3f7617996042938c727d4c3f7e1652ce8855fe270e826b9c1b281db"} Dec 10 14:53:47 crc kubenswrapper[4847]: I1210 14:53:47.283157 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" event={"ID":"7a627a8d-ac7d-4d5d-9258-3342ca59930d","Type":"ContainerStarted","Data":"6ed740bf4d6aefb51a9d0ffb12240ec6ca3e4498537636bf9140252b889aace0"} Dec 10 14:53:47 crc kubenswrapper[4847]: I1210 14:53:47.306078 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" podStartSLOduration=1.8397124630000001 podStartE2EDuration="2.306042151s" podCreationTimestamp="2025-12-10 14:53:45 +0000 UTC" firstStartedPulling="2025-12-10 14:53:46.2863361 +0000 UTC m=+1775.855553730" lastFinishedPulling="2025-12-10 14:53:46.752665788 +0000 UTC m=+1776.321883418" observedRunningTime="2025-12-10 14:53:47.30164748 +0000 UTC m=+1776.870865120" watchObservedRunningTime="2025-12-10 14:53:47.306042151 +0000 UTC m=+1776.875259781" Dec 10 14:53:47 crc kubenswrapper[4847]: I1210 14:53:47.760206 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:53:47 crc kubenswrapper[4847]: E1210 14:53:47.760796 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:53:52 crc kubenswrapper[4847]: I1210 14:53:52.333260 4847 generic.go:334] "Generic (PLEG): container finished" podID="7a627a8d-ac7d-4d5d-9258-3342ca59930d" containerID="0276f0c6b3f7617996042938c727d4c3f7e1652ce8855fe270e826b9c1b281db" exitCode=0 Dec 10 14:53:52 crc kubenswrapper[4847]: I1210 14:53:52.333393 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" event={"ID":"7a627a8d-ac7d-4d5d-9258-3342ca59930d","Type":"ContainerDied","Data":"0276f0c6b3f7617996042938c727d4c3f7e1652ce8855fe270e826b9c1b281db"} Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.793456 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.886793 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory\") pod \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.886856 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key\") pod \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.886890 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2bz2\" (UniqueName: \"kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2\") pod \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\" (UID: \"7a627a8d-ac7d-4d5d-9258-3342ca59930d\") " Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.915582 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2" (OuterVolumeSpecName: "kube-api-access-s2bz2") pod "7a627a8d-ac7d-4d5d-9258-3342ca59930d" (UID: "7a627a8d-ac7d-4d5d-9258-3342ca59930d"). InnerVolumeSpecName "kube-api-access-s2bz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.925967 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory" (OuterVolumeSpecName: "inventory") pod "7a627a8d-ac7d-4d5d-9258-3342ca59930d" (UID: "7a627a8d-ac7d-4d5d-9258-3342ca59930d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.947202 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a627a8d-ac7d-4d5d-9258-3342ca59930d" (UID: "7a627a8d-ac7d-4d5d-9258-3342ca59930d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.989090 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.989137 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a627a8d-ac7d-4d5d-9258-3342ca59930d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:53 crc kubenswrapper[4847]: I1210 14:53:53.989156 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2bz2\" (UniqueName: \"kubernetes.io/projected/7a627a8d-ac7d-4d5d-9258-3342ca59930d-kube-api-access-s2bz2\") on node \"crc\" DevicePath \"\"" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.355506 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" event={"ID":"7a627a8d-ac7d-4d5d-9258-3342ca59930d","Type":"ContainerDied","Data":"6ed740bf4d6aefb51a9d0ffb12240ec6ca3e4498537636bf9140252b889aace0"} Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.355566 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ed740bf4d6aefb51a9d0ffb12240ec6ca3e4498537636bf9140252b889aace0" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.355568 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.441445 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g"] Dec 10 14:53:54 crc kubenswrapper[4847]: E1210 14:53:54.441941 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a627a8d-ac7d-4d5d-9258-3342ca59930d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.441968 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a627a8d-ac7d-4d5d-9258-3342ca59930d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.442232 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a627a8d-ac7d-4d5d-9258-3342ca59930d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.443086 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.446516 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.446791 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.446949 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.449368 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.466620 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g"] Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.602000 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.602325 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spm4m\" (UniqueName: \"kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.602387 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.704294 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.704781 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spm4m\" (UniqueName: \"kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.704804 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.710377 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.710377 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.725034 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spm4m\" (UniqueName: \"kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-hvd6g\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:54 crc kubenswrapper[4847]: I1210 14:53:54.775279 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:53:55 crc kubenswrapper[4847]: I1210 14:53:55.386572 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g"] Dec 10 14:53:56 crc kubenswrapper[4847]: I1210 14:53:56.389485 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" event={"ID":"d4c9d846-3adb-4f21-8945-f27d14123b1e","Type":"ContainerStarted","Data":"3405c41947846c724fce1805f64eceafe8e64cd8cc7ca44143c1bedbb29e130b"} Dec 10 14:53:57 crc kubenswrapper[4847]: I1210 14:53:57.401782 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" event={"ID":"d4c9d846-3adb-4f21-8945-f27d14123b1e","Type":"ContainerStarted","Data":"454d90be21c375a0cee6a4469e1718c72e1460a2c6b12a895aecaa9a518fc9ed"} Dec 10 14:53:57 crc kubenswrapper[4847]: I1210 14:53:57.420926 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" podStartSLOduration=2.61299018 podStartE2EDuration="3.420899257s" podCreationTimestamp="2025-12-10 14:53:54 +0000 UTC" firstStartedPulling="2025-12-10 14:53:55.378982034 +0000 UTC m=+1784.948199664" lastFinishedPulling="2025-12-10 14:53:56.186891111 +0000 UTC m=+1785.756108741" observedRunningTime="2025-12-10 14:53:57.417201316 +0000 UTC m=+1786.986418966" watchObservedRunningTime="2025-12-10 14:53:57.420899257 +0000 UTC m=+1786.990116897" Dec 10 14:54:02 crc kubenswrapper[4847]: I1210 14:54:02.760680 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:54:02 crc kubenswrapper[4847]: E1210 14:54:02.761993 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:54:14 crc kubenswrapper[4847]: I1210 14:54:14.759409 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:54:14 crc kubenswrapper[4847]: E1210 14:54:14.760093 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:54:17 crc kubenswrapper[4847]: I1210 14:54:17.050198 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-fsqtk"] Dec 10 14:54:17 crc kubenswrapper[4847]: I1210 14:54:17.062842 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-fsqtk"] Dec 10 14:54:18 crc kubenswrapper[4847]: I1210 14:54:18.034553 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-64b8-account-create-update-6q5jl"] Dec 10 14:54:18 crc kubenswrapper[4847]: I1210 14:54:18.043579 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-64b8-account-create-update-6q5jl"] Dec 10 14:54:18 crc kubenswrapper[4847]: I1210 14:54:18.772670 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6" path="/var/lib/kubelet/pods/20ad03ce-c5f0-4ba1-8f44-bf7049dee1f6/volumes" Dec 10 14:54:18 crc kubenswrapper[4847]: I1210 14:54:18.775143 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992" path="/var/lib/kubelet/pods/b6c2e2c8-cc11-4db9-a572-4fb3e4e9c992/volumes" Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.037036 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bkdzb"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.047048 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9a76-account-create-update-6lfrc"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.057359 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9a76-account-create-update-6lfrc"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.068639 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bkdzb"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.080134 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-b8f8x"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.090833 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-d0de-account-create-update-k2qzx"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.103547 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-d0de-account-create-update-k2qzx"] Dec 10 14:54:19 crc kubenswrapper[4847]: I1210 14:54:19.115953 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-b8f8x"] Dec 10 14:54:20 crc kubenswrapper[4847]: I1210 14:54:20.776849 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193125a6-9a88-47b0-b462-daab3aa61691" path="/var/lib/kubelet/pods/193125a6-9a88-47b0-b462-daab3aa61691/volumes" Dec 10 14:54:20 crc kubenswrapper[4847]: I1210 14:54:20.778263 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333a1d69-c6b6-434e-b0e3-7c43c39a2e1f" path="/var/lib/kubelet/pods/333a1d69-c6b6-434e-b0e3-7c43c39a2e1f/volumes" Dec 10 14:54:20 crc kubenswrapper[4847]: I1210 14:54:20.779128 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a4fd885-5fc1-4bc8-9b47-57e3103e9197" path="/var/lib/kubelet/pods/7a4fd885-5fc1-4bc8-9b47-57e3103e9197/volumes" Dec 10 14:54:20 crc kubenswrapper[4847]: I1210 14:54:20.779862 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6a1839-419a-4f80-9c45-ff37701bb916" path="/var/lib/kubelet/pods/7b6a1839-419a-4f80-9c45-ff37701bb916/volumes" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.147834 4847 scope.go:117] "RemoveContainer" containerID="cf2984bedaf57c40bdb660c3c77aa01e206f83a112cda89b6919d727739a0bf2" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.173515 4847 scope.go:117] "RemoveContainer" containerID="3e334ccf765e497d8ace6e292f05aef56b61a77a92433f93f0064cc662088c75" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.255582 4847 scope.go:117] "RemoveContainer" containerID="f87b9390965af2b56a58dfb80fed954356e271af1804f0d17c8ea49bf6d6218b" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.312930 4847 scope.go:117] "RemoveContainer" containerID="b20880d06e12340b52aae3309e07883025acd94e4c8f029f673d432787736b68" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.356440 4847 scope.go:117] "RemoveContainer" containerID="b2238040e5778411991958e5030a4cf7b7ea1247928baeecb4b797a1fdabeeda" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.413383 4847 scope.go:117] "RemoveContainer" containerID="451ad0ad43f8dc82e00f2764d86a6ff59a5dabcc5536ca3c0a3f783a427cd911" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.498587 4847 scope.go:117] "RemoveContainer" containerID="27a23ddfa0d2cd6f6008fe3b9cd1f054779cb51aba9c84bab00027dfdf2bba16" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.523234 4847 scope.go:117] "RemoveContainer" containerID="2d86267a03d84cdfa47b024e8bd134de1fe886a05cdf39d7cb5a1106f42fed9e" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.556121 4847 scope.go:117] "RemoveContainer" containerID="8413752fb972c003cf7769c652ddc5c3ea73dd479cd96ce2070628b29d74a6cb" Dec 10 14:54:23 crc kubenswrapper[4847]: I1210 14:54:23.584310 4847 scope.go:117] "RemoveContainer" containerID="48c20f73a3a36a202fdba3cdad451a92ddee0e3509aabeba2d8450409fbd3942" Dec 10 14:54:25 crc kubenswrapper[4847]: I1210 14:54:25.760033 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:54:25 crc kubenswrapper[4847]: E1210 14:54:25.760636 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 14:54:32 crc kubenswrapper[4847]: I1210 14:54:32.763247 4847 generic.go:334] "Generic (PLEG): container finished" podID="d4c9d846-3adb-4f21-8945-f27d14123b1e" containerID="454d90be21c375a0cee6a4469e1718c72e1460a2c6b12a895aecaa9a518fc9ed" exitCode=0 Dec 10 14:54:32 crc kubenswrapper[4847]: I1210 14:54:32.769654 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" event={"ID":"d4c9d846-3adb-4f21-8945-f27d14123b1e","Type":"ContainerDied","Data":"454d90be21c375a0cee6a4469e1718c72e1460a2c6b12a895aecaa9a518fc9ed"} Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.220706 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.282544 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key\") pod \"d4c9d846-3adb-4f21-8945-f27d14123b1e\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.282730 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory\") pod \"d4c9d846-3adb-4f21-8945-f27d14123b1e\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.282820 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spm4m\" (UniqueName: \"kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m\") pod \"d4c9d846-3adb-4f21-8945-f27d14123b1e\" (UID: \"d4c9d846-3adb-4f21-8945-f27d14123b1e\") " Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.287990 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m" (OuterVolumeSpecName: "kube-api-access-spm4m") pod "d4c9d846-3adb-4f21-8945-f27d14123b1e" (UID: "d4c9d846-3adb-4f21-8945-f27d14123b1e"). InnerVolumeSpecName "kube-api-access-spm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.314801 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory" (OuterVolumeSpecName: "inventory") pod "d4c9d846-3adb-4f21-8945-f27d14123b1e" (UID: "d4c9d846-3adb-4f21-8945-f27d14123b1e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.315359 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4c9d846-3adb-4f21-8945-f27d14123b1e" (UID: "d4c9d846-3adb-4f21-8945-f27d14123b1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.385662 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.386118 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c9d846-3adb-4f21-8945-f27d14123b1e-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.386300 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spm4m\" (UniqueName: \"kubernetes.io/projected/d4c9d846-3adb-4f21-8945-f27d14123b1e-kube-api-access-spm4m\") on node \"crc\" DevicePath \"\"" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.787093 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" event={"ID":"d4c9d846-3adb-4f21-8945-f27d14123b1e","Type":"ContainerDied","Data":"3405c41947846c724fce1805f64eceafe8e64cd8cc7ca44143c1bedbb29e130b"} Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.787165 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3405c41947846c724fce1805f64eceafe8e64cd8cc7ca44143c1bedbb29e130b" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.787213 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-hvd6g" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.888885 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7"] Dec 10 14:54:34 crc kubenswrapper[4847]: E1210 14:54:34.889486 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c9d846-3adb-4f21-8945-f27d14123b1e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.889532 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c9d846-3adb-4f21-8945-f27d14123b1e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.889846 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c9d846-3adb-4f21-8945-f27d14123b1e" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.890764 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.892967 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.897286 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.897343 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.897864 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:54:34 crc kubenswrapper[4847]: I1210 14:54:34.900037 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7"] Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.000968 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb2dz\" (UniqueName: \"kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.001065 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.001119 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.103293 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.103427 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.103563 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb2dz\" (UniqueName: \"kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.111001 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.112924 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.124400 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb2dz\" (UniqueName: \"kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.212371 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.779362 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7"] Dec 10 14:54:35 crc kubenswrapper[4847]: I1210 14:54:35.800944 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" event={"ID":"da983cff-081c-4811-9ed3-f3cc109489bd","Type":"ContainerStarted","Data":"bb03d4dc8d1116a37c2aa39307c36d9d3920e3b6a80d0d3064c288fd702c7bad"} Dec 10 14:54:36 crc kubenswrapper[4847]: I1210 14:54:36.814373 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" event={"ID":"da983cff-081c-4811-9ed3-f3cc109489bd","Type":"ContainerStarted","Data":"ef475530caac8d3808587a5d99f2402a8d234d7128df2d1ce14d5135f987c6b5"} Dec 10 14:54:36 crc kubenswrapper[4847]: I1210 14:54:36.844090 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" podStartSLOduration=2.269415401 podStartE2EDuration="2.844049329s" podCreationTimestamp="2025-12-10 14:54:34 +0000 UTC" firstStartedPulling="2025-12-10 14:54:35.789947501 +0000 UTC m=+1825.359165131" lastFinishedPulling="2025-12-10 14:54:36.364581429 +0000 UTC m=+1825.933799059" observedRunningTime="2025-12-10 14:54:36.83939131 +0000 UTC m=+1826.408608960" watchObservedRunningTime="2025-12-10 14:54:36.844049329 +0000 UTC m=+1826.413266959" Dec 10 14:54:39 crc kubenswrapper[4847]: I1210 14:54:39.760000 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:54:40 crc kubenswrapper[4847]: I1210 14:54:40.861385 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88"} Dec 10 14:54:48 crc kubenswrapper[4847]: I1210 14:54:48.044671 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lgckt"] Dec 10 14:54:48 crc kubenswrapper[4847]: I1210 14:54:48.053222 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-lgckt"] Dec 10 14:54:48 crc kubenswrapper[4847]: I1210 14:54:48.773467 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f3fb84-6bbc-478c-9f5c-a6adcd0b1908" path="/var/lib/kubelet/pods/56f3fb84-6bbc-478c-9f5c-a6adcd0b1908/volumes" Dec 10 14:55:07 crc kubenswrapper[4847]: I1210 14:55:07.054671 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zdcv2"] Dec 10 14:55:07 crc kubenswrapper[4847]: I1210 14:55:07.065684 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zdcv2"] Dec 10 14:55:08 crc kubenswrapper[4847]: I1210 14:55:08.771040 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dd1e94d-67a9-443c-9870-81a88bf56445" path="/var/lib/kubelet/pods/5dd1e94d-67a9-443c-9870-81a88bf56445/volumes" Dec 10 14:55:10 crc kubenswrapper[4847]: I1210 14:55:10.045067 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8bgjb"] Dec 10 14:55:10 crc kubenswrapper[4847]: I1210 14:55:10.058348 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8bgjb"] Dec 10 14:55:10 crc kubenswrapper[4847]: I1210 14:55:10.779214 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0ae39e9-b424-481b-a316-403a59eb4837" path="/var/lib/kubelet/pods/f0ae39e9-b424-481b-a316-403a59eb4837/volumes" Dec 10 14:55:23 crc kubenswrapper[4847]: I1210 14:55:23.798887 4847 scope.go:117] "RemoveContainer" containerID="68a84f74c916db4f5a2cbd1b3d639386df8b56b0b4509c4be4dde8a49eba44cf" Dec 10 14:55:23 crc kubenswrapper[4847]: I1210 14:55:23.862306 4847 scope.go:117] "RemoveContainer" containerID="cd7aebf5817a579950b349a82d3860e10ad53f542e61f2fd453d39385993045d" Dec 10 14:55:23 crc kubenswrapper[4847]: I1210 14:55:23.960651 4847 scope.go:117] "RemoveContainer" containerID="4f0d3ca62098db3f83d740b5799f740e768d4f35515a52f4dbc37dbe4955d797" Dec 10 14:55:24 crc kubenswrapper[4847]: I1210 14:55:24.241647 4847 generic.go:334] "Generic (PLEG): container finished" podID="da983cff-081c-4811-9ed3-f3cc109489bd" containerID="ef475530caac8d3808587a5d99f2402a8d234d7128df2d1ce14d5135f987c6b5" exitCode=0 Dec 10 14:55:24 crc kubenswrapper[4847]: I1210 14:55:24.241744 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" event={"ID":"da983cff-081c-4811-9ed3-f3cc109489bd","Type":"ContainerDied","Data":"ef475530caac8d3808587a5d99f2402a8d234d7128df2d1ce14d5135f987c6b5"} Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.677591 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.793922 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb2dz\" (UniqueName: \"kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz\") pod \"da983cff-081c-4811-9ed3-f3cc109489bd\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.794076 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key\") pod \"da983cff-081c-4811-9ed3-f3cc109489bd\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.794154 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory\") pod \"da983cff-081c-4811-9ed3-f3cc109489bd\" (UID: \"da983cff-081c-4811-9ed3-f3cc109489bd\") " Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.800227 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz" (OuterVolumeSpecName: "kube-api-access-nb2dz") pod "da983cff-081c-4811-9ed3-f3cc109489bd" (UID: "da983cff-081c-4811-9ed3-f3cc109489bd"). InnerVolumeSpecName "kube-api-access-nb2dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.822804 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory" (OuterVolumeSpecName: "inventory") pod "da983cff-081c-4811-9ed3-f3cc109489bd" (UID: "da983cff-081c-4811-9ed3-f3cc109489bd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.824889 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da983cff-081c-4811-9ed3-f3cc109489bd" (UID: "da983cff-081c-4811-9ed3-f3cc109489bd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.897225 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb2dz\" (UniqueName: \"kubernetes.io/projected/da983cff-081c-4811-9ed3-f3cc109489bd-kube-api-access-nb2dz\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.897264 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:25 crc kubenswrapper[4847]: I1210 14:55:25.897275 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da983cff-081c-4811-9ed3-f3cc109489bd-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.261070 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" event={"ID":"da983cff-081c-4811-9ed3-f3cc109489bd","Type":"ContainerDied","Data":"bb03d4dc8d1116a37c2aa39307c36d9d3920e3b6a80d0d3064c288fd702c7bad"} Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.261368 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb03d4dc8d1116a37c2aa39307c36d9d3920e3b6a80d0d3064c288fd702c7bad" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.261122 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.355192 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4ckkd"] Dec 10 14:55:26 crc kubenswrapper[4847]: E1210 14:55:26.355698 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da983cff-081c-4811-9ed3-f3cc109489bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.355748 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="da983cff-081c-4811-9ed3-f3cc109489bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.355989 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="da983cff-081c-4811-9ed3-f3cc109489bd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.356635 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.364226 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.364405 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.364476 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.366452 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.369203 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4ckkd"] Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.509135 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.509229 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.509411 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rttv7\" (UniqueName: \"kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.610992 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rttv7\" (UniqueName: \"kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.611047 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.611104 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.617119 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.617668 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.628031 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rttv7\" (UniqueName: \"kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7\") pod \"ssh-known-hosts-edpm-deployment-4ckkd\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:26 crc kubenswrapper[4847]: I1210 14:55:26.697375 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:27 crc kubenswrapper[4847]: I1210 14:55:27.270859 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4ckkd"] Dec 10 14:55:28 crc kubenswrapper[4847]: I1210 14:55:28.280545 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" event={"ID":"6e914d68-6bc2-420a-b0cb-3f424d71652c","Type":"ContainerStarted","Data":"59ad6b7c8b8c3190a7e2fd52ca28d3e0ae521723c8ac441a2a8e515ae45d535d"} Dec 10 14:55:31 crc kubenswrapper[4847]: I1210 14:55:31.309065 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" event={"ID":"6e914d68-6bc2-420a-b0cb-3f424d71652c","Type":"ContainerStarted","Data":"a2a86f1d4bd9a2ed96debf1cc81e81cb72b9cd849421e08552253131574cd9c1"} Dec 10 14:55:31 crc kubenswrapper[4847]: I1210 14:55:31.329140 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" podStartSLOduration=2.40833211 podStartE2EDuration="5.3291182s" podCreationTimestamp="2025-12-10 14:55:26 +0000 UTC" firstStartedPulling="2025-12-10 14:55:27.278466398 +0000 UTC m=+1876.847684028" lastFinishedPulling="2025-12-10 14:55:30.199252488 +0000 UTC m=+1879.768470118" observedRunningTime="2025-12-10 14:55:31.324621057 +0000 UTC m=+1880.893838707" watchObservedRunningTime="2025-12-10 14:55:31.3291182 +0000 UTC m=+1880.898335830" Dec 10 14:55:38 crc kubenswrapper[4847]: I1210 14:55:38.369673 4847 generic.go:334] "Generic (PLEG): container finished" podID="6e914d68-6bc2-420a-b0cb-3f424d71652c" containerID="a2a86f1d4bd9a2ed96debf1cc81e81cb72b9cd849421e08552253131574cd9c1" exitCode=0 Dec 10 14:55:38 crc kubenswrapper[4847]: I1210 14:55:38.369896 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" event={"ID":"6e914d68-6bc2-420a-b0cb-3f424d71652c","Type":"ContainerDied","Data":"a2a86f1d4bd9a2ed96debf1cc81e81cb72b9cd849421e08552253131574cd9c1"} Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.839001 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.868330 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rttv7\" (UniqueName: \"kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7\") pod \"6e914d68-6bc2-420a-b0cb-3f424d71652c\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.868497 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0\") pod \"6e914d68-6bc2-420a-b0cb-3f424d71652c\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.868669 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam\") pod \"6e914d68-6bc2-420a-b0cb-3f424d71652c\" (UID: \"6e914d68-6bc2-420a-b0cb-3f424d71652c\") " Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.889473 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7" (OuterVolumeSpecName: "kube-api-access-rttv7") pod "6e914d68-6bc2-420a-b0cb-3f424d71652c" (UID: "6e914d68-6bc2-420a-b0cb-3f424d71652c"). InnerVolumeSpecName "kube-api-access-rttv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.902341 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6e914d68-6bc2-420a-b0cb-3f424d71652c" (UID: "6e914d68-6bc2-420a-b0cb-3f424d71652c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.904196 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6e914d68-6bc2-420a-b0cb-3f424d71652c" (UID: "6e914d68-6bc2-420a-b0cb-3f424d71652c"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.972191 4847 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.972248 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6e914d68-6bc2-420a-b0cb-3f424d71652c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:39 crc kubenswrapper[4847]: I1210 14:55:39.972261 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rttv7\" (UniqueName: \"kubernetes.io/projected/6e914d68-6bc2-420a-b0cb-3f424d71652c-kube-api-access-rttv7\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.395393 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" event={"ID":"6e914d68-6bc2-420a-b0cb-3f424d71652c","Type":"ContainerDied","Data":"59ad6b7c8b8c3190a7e2fd52ca28d3e0ae521723c8ac441a2a8e515ae45d535d"} Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.395466 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59ad6b7c8b8c3190a7e2fd52ca28d3e0ae521723c8ac441a2a8e515ae45d535d" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.395464 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4ckkd" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.553123 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g"] Dec 10 14:55:40 crc kubenswrapper[4847]: E1210 14:55:40.553848 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e914d68-6bc2-420a-b0cb-3f424d71652c" containerName="ssh-known-hosts-edpm-deployment" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.553878 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e914d68-6bc2-420a-b0cb-3f424d71652c" containerName="ssh-known-hosts-edpm-deployment" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.554158 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e914d68-6bc2-420a-b0cb-3f424d71652c" containerName="ssh-known-hosts-edpm-deployment" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.555097 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.558403 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.558665 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.558872 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.559101 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.566829 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g"] Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.589694 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2wlv\" (UniqueName: \"kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.589801 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.590006 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.692516 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.692617 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2wlv\" (UniqueName: \"kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.692678 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.697329 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.697536 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.713355 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2wlv\" (UniqueName: \"kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-xs56g\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:40 crc kubenswrapper[4847]: I1210 14:55:40.898332 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:41 crc kubenswrapper[4847]: I1210 14:55:41.534025 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g"] Dec 10 14:55:42 crc kubenswrapper[4847]: I1210 14:55:42.417096 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" event={"ID":"f9991c0e-7804-406a-bc24-92c5d5d7fa61","Type":"ContainerStarted","Data":"4b7c7df9d2e99651265ece63726eff956872951fcc9e6c34f159e6482a623a78"} Dec 10 14:55:42 crc kubenswrapper[4847]: I1210 14:55:42.417669 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" event={"ID":"f9991c0e-7804-406a-bc24-92c5d5d7fa61","Type":"ContainerStarted","Data":"5ddf72cd7b13f1c159072213addabd8141063d90490a2f08d36901199e7f2cfa"} Dec 10 14:55:43 crc kubenswrapper[4847]: I1210 14:55:43.468535 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" podStartSLOduration=2.923469761 podStartE2EDuration="3.46850477s" podCreationTimestamp="2025-12-10 14:55:40 +0000 UTC" firstStartedPulling="2025-12-10 14:55:41.544626619 +0000 UTC m=+1891.113844249" lastFinishedPulling="2025-12-10 14:55:42.089661628 +0000 UTC m=+1891.658879258" observedRunningTime="2025-12-10 14:55:43.459476491 +0000 UTC m=+1893.028694131" watchObservedRunningTime="2025-12-10 14:55:43.46850477 +0000 UTC m=+1893.037722400" Dec 10 14:55:50 crc kubenswrapper[4847]: I1210 14:55:50.506003 4847 generic.go:334] "Generic (PLEG): container finished" podID="f9991c0e-7804-406a-bc24-92c5d5d7fa61" containerID="4b7c7df9d2e99651265ece63726eff956872951fcc9e6c34f159e6482a623a78" exitCode=0 Dec 10 14:55:50 crc kubenswrapper[4847]: I1210 14:55:50.506084 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" event={"ID":"f9991c0e-7804-406a-bc24-92c5d5d7fa61","Type":"ContainerDied","Data":"4b7c7df9d2e99651265ece63726eff956872951fcc9e6c34f159e6482a623a78"} Dec 10 14:55:51 crc kubenswrapper[4847]: I1210 14:55:51.048372 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pwcls"] Dec 10 14:55:51 crc kubenswrapper[4847]: I1210 14:55:51.057537 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pwcls"] Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.143021 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.237741 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory\") pod \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.238148 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2wlv\" (UniqueName: \"kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv\") pod \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.238253 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key\") pod \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\" (UID: \"f9991c0e-7804-406a-bc24-92c5d5d7fa61\") " Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.250239 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv" (OuterVolumeSpecName: "kube-api-access-b2wlv") pod "f9991c0e-7804-406a-bc24-92c5d5d7fa61" (UID: "f9991c0e-7804-406a-bc24-92c5d5d7fa61"). InnerVolumeSpecName "kube-api-access-b2wlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.271181 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory" (OuterVolumeSpecName: "inventory") pod "f9991c0e-7804-406a-bc24-92c5d5d7fa61" (UID: "f9991c0e-7804-406a-bc24-92c5d5d7fa61"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.276676 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9991c0e-7804-406a-bc24-92c5d5d7fa61" (UID: "f9991c0e-7804-406a-bc24-92c5d5d7fa61"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.340173 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.340235 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2wlv\" (UniqueName: \"kubernetes.io/projected/f9991c0e-7804-406a-bc24-92c5d5d7fa61-kube-api-access-b2wlv\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.340248 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9991c0e-7804-406a-bc24-92c5d5d7fa61-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.693216 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" event={"ID":"f9991c0e-7804-406a-bc24-92c5d5d7fa61","Type":"ContainerDied","Data":"5ddf72cd7b13f1c159072213addabd8141063d90490a2f08d36901199e7f2cfa"} Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.693289 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ddf72cd7b13f1c159072213addabd8141063d90490a2f08d36901199e7f2cfa" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.693323 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-xs56g" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.747503 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr"] Dec 10 14:55:52 crc kubenswrapper[4847]: E1210 14:55:52.748019 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9991c0e-7804-406a-bc24-92c5d5d7fa61" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.748051 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9991c0e-7804-406a-bc24-92c5d5d7fa61" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.748333 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9991c0e-7804-406a-bc24-92c5d5d7fa61" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.749334 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.751872 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.752108 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.752149 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.755261 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.788307 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb71e2f0-928f-4d58-adf9-c5127aa279c5" path="/var/lib/kubelet/pods/bb71e2f0-928f-4d58-adf9-c5127aa279c5/volumes" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.788985 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr"] Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.850466 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.850572 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.850790 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgwph\" (UniqueName: \"kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.952634 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgwph\" (UniqueName: \"kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.952862 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.952896 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.957480 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.987498 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:52 crc kubenswrapper[4847]: I1210 14:55:52.995099 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgwph\" (UniqueName: \"kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:53 crc kubenswrapper[4847]: I1210 14:55:53.095624 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:55:53 crc kubenswrapper[4847]: I1210 14:55:53.735379 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr"] Dec 10 14:55:54 crc kubenswrapper[4847]: I1210 14:55:54.722040 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" event={"ID":"4ef271cb-44e6-4f26-95fc-3ca47f6ab290","Type":"ContainerStarted","Data":"7f7c0afce488efc445c6629242762d1bedb7f6e3fefcfb6795dd105cceb58a48"} Dec 10 14:55:54 crc kubenswrapper[4847]: I1210 14:55:54.723227 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" event={"ID":"4ef271cb-44e6-4f26-95fc-3ca47f6ab290","Type":"ContainerStarted","Data":"075ba058ab63bff772d3d9f8b6bff402304cfda6777118bcd17f35465f828cc0"} Dec 10 14:55:54 crc kubenswrapper[4847]: I1210 14:55:54.748980 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" podStartSLOduration=2.22405812 podStartE2EDuration="2.748951785s" podCreationTimestamp="2025-12-10 14:55:52 +0000 UTC" firstStartedPulling="2025-12-10 14:55:53.752422239 +0000 UTC m=+1903.321639869" lastFinishedPulling="2025-12-10 14:55:54.277315904 +0000 UTC m=+1903.846533534" observedRunningTime="2025-12-10 14:55:54.738842308 +0000 UTC m=+1904.308059948" watchObservedRunningTime="2025-12-10 14:55:54.748951785 +0000 UTC m=+1904.318169415" Dec 10 14:56:03 crc kubenswrapper[4847]: I1210 14:56:03.831547 4847 generic.go:334] "Generic (PLEG): container finished" podID="4ef271cb-44e6-4f26-95fc-3ca47f6ab290" containerID="7f7c0afce488efc445c6629242762d1bedb7f6e3fefcfb6795dd105cceb58a48" exitCode=0 Dec 10 14:56:03 crc kubenswrapper[4847]: I1210 14:56:03.831594 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" event={"ID":"4ef271cb-44e6-4f26-95fc-3ca47f6ab290","Type":"ContainerDied","Data":"7f7c0afce488efc445c6629242762d1bedb7f6e3fefcfb6795dd105cceb58a48"} Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.279983 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.418667 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key\") pod \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.418738 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory\") pod \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.418854 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgwph\" (UniqueName: \"kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph\") pod \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\" (UID: \"4ef271cb-44e6-4f26-95fc-3ca47f6ab290\") " Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.453084 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph" (OuterVolumeSpecName: "kube-api-access-cgwph") pod "4ef271cb-44e6-4f26-95fc-3ca47f6ab290" (UID: "4ef271cb-44e6-4f26-95fc-3ca47f6ab290"). InnerVolumeSpecName "kube-api-access-cgwph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.456935 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory" (OuterVolumeSpecName: "inventory") pod "4ef271cb-44e6-4f26-95fc-3ca47f6ab290" (UID: "4ef271cb-44e6-4f26-95fc-3ca47f6ab290"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.457603 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ef271cb-44e6-4f26-95fc-3ca47f6ab290" (UID: "4ef271cb-44e6-4f26-95fc-3ca47f6ab290"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.521730 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.521781 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.521793 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgwph\" (UniqueName: \"kubernetes.io/projected/4ef271cb-44e6-4f26-95fc-3ca47f6ab290-kube-api-access-cgwph\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.851851 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" event={"ID":"4ef271cb-44e6-4f26-95fc-3ca47f6ab290","Type":"ContainerDied","Data":"075ba058ab63bff772d3d9f8b6bff402304cfda6777118bcd17f35465f828cc0"} Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.851902 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.851902 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="075ba058ab63bff772d3d9f8b6bff402304cfda6777118bcd17f35465f828cc0" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.927406 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv"] Dec 10 14:56:05 crc kubenswrapper[4847]: E1210 14:56:05.927874 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef271cb-44e6-4f26-95fc-3ca47f6ab290" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.927891 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef271cb-44e6-4f26-95fc-3ca47f6ab290" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.928186 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef271cb-44e6-4f26-95fc-3ca47f6ab290" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.928856 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.931072 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.931463 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.931783 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.934343 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.934889 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.935030 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.935385 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.935517 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:56:05 crc kubenswrapper[4847]: I1210 14:56:05.952843 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv"] Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032019 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032335 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032358 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032377 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032397 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032415 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032439 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032498 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkkzg\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032531 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032556 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032611 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032705 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032744 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.032769 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134700 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134772 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134792 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134812 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134832 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134851 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134873 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134932 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkkzg\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134968 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.134996 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.135815 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.135914 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.136037 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.136736 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.140028 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.141074 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.141686 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.141752 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.142128 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.142366 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.143160 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.143815 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.144196 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.144761 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.145218 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.145967 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.146025 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.157189 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkkzg\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jlctv\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.256887 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:06 crc kubenswrapper[4847]: I1210 14:56:06.852482 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv"] Dec 10 14:56:07 crc kubenswrapper[4847]: I1210 14:56:07.871323 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" event={"ID":"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1","Type":"ContainerStarted","Data":"8ea4a40105ee5433747fce5913f7fc23e07763a4a9e6ffef2d055020a606496e"} Dec 10 14:56:07 crc kubenswrapper[4847]: I1210 14:56:07.872344 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" event={"ID":"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1","Type":"ContainerStarted","Data":"28c415406c6ea1a48eed5f32ef7db3de4a73267a4a06be207713134195f08781"} Dec 10 14:56:07 crc kubenswrapper[4847]: I1210 14:56:07.898281 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" podStartSLOduration=2.400669726 podStartE2EDuration="2.898256782s" podCreationTimestamp="2025-12-10 14:56:05 +0000 UTC" firstStartedPulling="2025-12-10 14:56:06.869670833 +0000 UTC m=+1916.438888463" lastFinishedPulling="2025-12-10 14:56:07.367257889 +0000 UTC m=+1916.936475519" observedRunningTime="2025-12-10 14:56:07.893617505 +0000 UTC m=+1917.462835125" watchObservedRunningTime="2025-12-10 14:56:07.898256782 +0000 UTC m=+1917.467474412" Dec 10 14:56:24 crc kubenswrapper[4847]: I1210 14:56:24.080494 4847 scope.go:117] "RemoveContainer" containerID="098cffea187b81968549c2a1ae43d92cb3fc93d790fecaa25464ab888776a178" Dec 10 14:56:43 crc kubenswrapper[4847]: I1210 14:56:43.210577 4847 generic.go:334] "Generic (PLEG): container finished" podID="a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" containerID="8ea4a40105ee5433747fce5913f7fc23e07763a4a9e6ffef2d055020a606496e" exitCode=0 Dec 10 14:56:43 crc kubenswrapper[4847]: I1210 14:56:43.210670 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" event={"ID":"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1","Type":"ContainerDied","Data":"8ea4a40105ee5433747fce5913f7fc23e07763a4a9e6ffef2d055020a606496e"} Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.652712 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803682 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803781 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803813 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803869 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803913 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.803976 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804004 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804062 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804224 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804301 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804324 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804372 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkkzg\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804403 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.804441 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle\") pod \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\" (UID: \"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1\") " Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.812594 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.813660 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.813802 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.814049 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.815104 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.815917 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.816007 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.816047 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.817153 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.817703 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.817747 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg" (OuterVolumeSpecName: "kube-api-access-wkkzg") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "kube-api-access-wkkzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.818728 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.843055 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.845327 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory" (OuterVolumeSpecName: "inventory") pod "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" (UID: "a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909135 4847 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909195 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909208 4847 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909223 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909235 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909248 4847 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909258 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909270 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkkzg\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-kube-api-access-wkkzg\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909280 4847 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909293 4847 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909306 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909314 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909328 4847 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:44 crc kubenswrapper[4847]: I1210 14:56:44.909340 4847 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.230009 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" event={"ID":"a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1","Type":"ContainerDied","Data":"28c415406c6ea1a48eed5f32ef7db3de4a73267a4a06be207713134195f08781"} Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.230074 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jlctv" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.230079 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c415406c6ea1a48eed5f32ef7db3de4a73267a4a06be207713134195f08781" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.345835 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf"] Dec 10 14:56:45 crc kubenswrapper[4847]: E1210 14:56:45.346211 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.346230 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.346441 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.347373 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.354627 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.354956 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.355322 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.355363 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.355518 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.364190 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf"] Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.522866 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.523253 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.523560 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6fd\" (UniqueName: \"kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.523729 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.523875 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.626706 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.627050 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.627170 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.627249 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6fd\" (UniqueName: \"kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.627305 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.627836 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.633532 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.634112 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.637091 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.646296 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6fd\" (UniqueName: \"kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vsngf\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:45 crc kubenswrapper[4847]: I1210 14:56:45.666649 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:56:46 crc kubenswrapper[4847]: I1210 14:56:46.206596 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf"] Dec 10 14:56:46 crc kubenswrapper[4847]: I1210 14:56:46.238940 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" event={"ID":"dfed455d-08ba-417a-975b-09dc84eaec54","Type":"ContainerStarted","Data":"4fe21db2d376fb18f8286bae328a9bb2dbd0bf9093b136873193ee56a1181c8d"} Dec 10 14:56:47 crc kubenswrapper[4847]: I1210 14:56:47.248191 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" event={"ID":"dfed455d-08ba-417a-975b-09dc84eaec54","Type":"ContainerStarted","Data":"6aff4ba2c242b4e694527efe1f45c14bce7dc80a16f783ee211ec863005bfa93"} Dec 10 14:56:47 crc kubenswrapper[4847]: I1210 14:56:47.273898 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" podStartSLOduration=1.614763414 podStartE2EDuration="2.273867382s" podCreationTimestamp="2025-12-10 14:56:45 +0000 UTC" firstStartedPulling="2025-12-10 14:56:46.219113632 +0000 UTC m=+1955.788331262" lastFinishedPulling="2025-12-10 14:56:46.8782176 +0000 UTC m=+1956.447435230" observedRunningTime="2025-12-10 14:56:47.261604685 +0000 UTC m=+1956.830822305" watchObservedRunningTime="2025-12-10 14:56:47.273867382 +0000 UTC m=+1956.843085022" Dec 10 14:57:01 crc kubenswrapper[4847]: I1210 14:57:01.011601 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:57:01 crc kubenswrapper[4847]: I1210 14:57:01.012277 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:57:31 crc kubenswrapper[4847]: I1210 14:57:31.011617 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:57:31 crc kubenswrapper[4847]: I1210 14:57:31.012756 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:57:45 crc kubenswrapper[4847]: I1210 14:57:45.780884 4847 generic.go:334] "Generic (PLEG): container finished" podID="dfed455d-08ba-417a-975b-09dc84eaec54" containerID="6aff4ba2c242b4e694527efe1f45c14bce7dc80a16f783ee211ec863005bfa93" exitCode=0 Dec 10 14:57:45 crc kubenswrapper[4847]: I1210 14:57:45.781012 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" event={"ID":"dfed455d-08ba-417a-975b-09dc84eaec54","Type":"ContainerDied","Data":"6aff4ba2c242b4e694527efe1f45c14bce7dc80a16f783ee211ec863005bfa93"} Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.253906 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.287667 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle\") pod \"dfed455d-08ba-417a-975b-09dc84eaec54\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.287833 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key\") pod \"dfed455d-08ba-417a-975b-09dc84eaec54\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.287897 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0\") pod \"dfed455d-08ba-417a-975b-09dc84eaec54\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.287988 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n6fd\" (UniqueName: \"kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd\") pod \"dfed455d-08ba-417a-975b-09dc84eaec54\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.288044 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory\") pod \"dfed455d-08ba-417a-975b-09dc84eaec54\" (UID: \"dfed455d-08ba-417a-975b-09dc84eaec54\") " Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.301523 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd" (OuterVolumeSpecName: "kube-api-access-5n6fd") pod "dfed455d-08ba-417a-975b-09dc84eaec54" (UID: "dfed455d-08ba-417a-975b-09dc84eaec54"). InnerVolumeSpecName "kube-api-access-5n6fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.305266 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dfed455d-08ba-417a-975b-09dc84eaec54" (UID: "dfed455d-08ba-417a-975b-09dc84eaec54"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.314774 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "dfed455d-08ba-417a-975b-09dc84eaec54" (UID: "dfed455d-08ba-417a-975b-09dc84eaec54"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.324976 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dfed455d-08ba-417a-975b-09dc84eaec54" (UID: "dfed455d-08ba-417a-975b-09dc84eaec54"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.325229 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory" (OuterVolumeSpecName: "inventory") pod "dfed455d-08ba-417a-975b-09dc84eaec54" (UID: "dfed455d-08ba-417a-975b-09dc84eaec54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.390290 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.390339 4847 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfed455d-08ba-417a-975b-09dc84eaec54-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.390351 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n6fd\" (UniqueName: \"kubernetes.io/projected/dfed455d-08ba-417a-975b-09dc84eaec54-kube-api-access-5n6fd\") on node \"crc\" DevicePath \"\"" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.390361 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.390370 4847 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfed455d-08ba-417a-975b-09dc84eaec54-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.806768 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" event={"ID":"dfed455d-08ba-417a-975b-09dc84eaec54","Type":"ContainerDied","Data":"4fe21db2d376fb18f8286bae328a9bb2dbd0bf9093b136873193ee56a1181c8d"} Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.807146 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fe21db2d376fb18f8286bae328a9bb2dbd0bf9093b136873193ee56a1181c8d" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.806933 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vsngf" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.937513 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c"] Dec 10 14:57:47 crc kubenswrapper[4847]: E1210 14:57:47.938159 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfed455d-08ba-417a-975b-09dc84eaec54" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.938176 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfed455d-08ba-417a-975b-09dc84eaec54" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.938510 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfed455d-08ba-417a-975b-09dc84eaec54" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.939325 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.942785 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.943062 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.943195 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.943382 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.944348 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.944858 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:57:47 crc kubenswrapper[4847]: I1210 14:57:47.958259 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c"] Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.001006 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.001422 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.001522 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.003060 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfdkk\" (UniqueName: \"kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.003148 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.003220 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105631 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfdkk\" (UniqueName: \"kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105689 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105734 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105755 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105833 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.105860 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.111064 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.111097 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.112336 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.118424 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.118940 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.122762 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfdkk\" (UniqueName: \"kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.263503 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.797371 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c"] Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.807956 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 14:57:48 crc kubenswrapper[4847]: I1210 14:57:48.816931 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" event={"ID":"04d0cfc8-2361-4e74-bacc-379a98054d97","Type":"ContainerStarted","Data":"8a5b682a77e752f5b09df6096fa749f0f2425e271ac44c06a0c24b11d58ece00"} Dec 10 14:57:49 crc kubenswrapper[4847]: I1210 14:57:49.825551 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" event={"ID":"04d0cfc8-2361-4e74-bacc-379a98054d97","Type":"ContainerStarted","Data":"437d806ae93079d0b53399807cfae6a3c71fd119b98dca1a0b3d4ef27c1d5ae4"} Dec 10 14:57:49 crc kubenswrapper[4847]: I1210 14:57:49.847896 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" podStartSLOduration=2.166471192 podStartE2EDuration="2.847877394s" podCreationTimestamp="2025-12-10 14:57:47 +0000 UTC" firstStartedPulling="2025-12-10 14:57:48.807615143 +0000 UTC m=+2018.376832773" lastFinishedPulling="2025-12-10 14:57:49.489021345 +0000 UTC m=+2019.058238975" observedRunningTime="2025-12-10 14:57:49.84701592 +0000 UTC m=+2019.416233540" watchObservedRunningTime="2025-12-10 14:57:49.847877394 +0000 UTC m=+2019.417095024" Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.011225 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.011842 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.011893 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.012639 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.012694 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88" gracePeriod=600 Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.941796 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88" exitCode=0 Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.942137 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88"} Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.942164 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d"} Dec 10 14:58:01 crc kubenswrapper[4847]: I1210 14:58:01.942179 4847 scope.go:117] "RemoveContainer" containerID="f779c1d66eeedbd8058d12e1b631f15370f411390ce231136d5a432ec49cd88e" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.612975 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.618444 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.625919 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.727280 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.727441 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.727654 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md84p\" (UniqueName: \"kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.830265 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.830738 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md84p\" (UniqueName: \"kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.830804 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.830902 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.831261 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.858110 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md84p\" (UniqueName: \"kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p\") pod \"redhat-operators-j9qsn\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:27 crc kubenswrapper[4847]: I1210 14:58:27.943898 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:28 crc kubenswrapper[4847]: I1210 14:58:28.445224 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:29 crc kubenswrapper[4847]: I1210 14:58:29.202091 4847 generic.go:334] "Generic (PLEG): container finished" podID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerID="d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0" exitCode=0 Dec 10 14:58:29 crc kubenswrapper[4847]: I1210 14:58:29.202640 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerDied","Data":"d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0"} Dec 10 14:58:29 crc kubenswrapper[4847]: I1210 14:58:29.202693 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerStarted","Data":"05fc228d5a4ac6e5d6ff39ddd3fa5f02a3a127e304ddec042cdb734da3846105"} Dec 10 14:58:31 crc kubenswrapper[4847]: I1210 14:58:31.223601 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerStarted","Data":"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de"} Dec 10 14:58:33 crc kubenswrapper[4847]: I1210 14:58:33.258369 4847 generic.go:334] "Generic (PLEG): container finished" podID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerID="e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de" exitCode=0 Dec 10 14:58:33 crc kubenswrapper[4847]: I1210 14:58:33.258448 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerDied","Data":"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de"} Dec 10 14:58:35 crc kubenswrapper[4847]: I1210 14:58:35.279244 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerStarted","Data":"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f"} Dec 10 14:58:35 crc kubenswrapper[4847]: I1210 14:58:35.304636 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9qsn" podStartSLOduration=3.043976908 podStartE2EDuration="8.304616262s" podCreationTimestamp="2025-12-10 14:58:27 +0000 UTC" firstStartedPulling="2025-12-10 14:58:29.207123443 +0000 UTC m=+2058.776341073" lastFinishedPulling="2025-12-10 14:58:34.467762807 +0000 UTC m=+2064.036980427" observedRunningTime="2025-12-10 14:58:35.297230439 +0000 UTC m=+2064.866448079" watchObservedRunningTime="2025-12-10 14:58:35.304616262 +0000 UTC m=+2064.873833882" Dec 10 14:58:36 crc kubenswrapper[4847]: I1210 14:58:36.289398 4847 generic.go:334] "Generic (PLEG): container finished" podID="04d0cfc8-2361-4e74-bacc-379a98054d97" containerID="437d806ae93079d0b53399807cfae6a3c71fd119b98dca1a0b3d4ef27c1d5ae4" exitCode=0 Dec 10 14:58:36 crc kubenswrapper[4847]: I1210 14:58:36.289458 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" event={"ID":"04d0cfc8-2361-4e74-bacc-379a98054d97","Type":"ContainerDied","Data":"437d806ae93079d0b53399807cfae6a3c71fd119b98dca1a0b3d4ef27c1d5ae4"} Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.791145 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940126 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940221 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940382 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940457 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940495 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfdkk\" (UniqueName: \"kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.940539 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key\") pod \"04d0cfc8-2361-4e74-bacc-379a98054d97\" (UID: \"04d0cfc8-2361-4e74-bacc-379a98054d97\") " Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.944926 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.946509 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.951201 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk" (OuterVolumeSpecName: "kube-api-access-tfdkk") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "kube-api-access-tfdkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.955033 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.978380 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.981218 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.981348 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:58:37 crc kubenswrapper[4847]: I1210 14:58:37.983116 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory" (OuterVolumeSpecName: "inventory") pod "04d0cfc8-2361-4e74-bacc-379a98054d97" (UID: "04d0cfc8-2361-4e74-bacc-379a98054d97"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.042823 4847 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.043171 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfdkk\" (UniqueName: \"kubernetes.io/projected/04d0cfc8-2361-4e74-bacc-379a98054d97-kube-api-access-tfdkk\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.043186 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.043199 4847 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.043211 4847 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.043223 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04d0cfc8-2361-4e74-bacc-379a98054d97-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.308346 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.308417 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c" event={"ID":"04d0cfc8-2361-4e74-bacc-379a98054d97","Type":"ContainerDied","Data":"8a5b682a77e752f5b09df6096fa749f0f2425e271ac44c06a0c24b11d58ece00"} Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.308455 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a5b682a77e752f5b09df6096fa749f0f2425e271ac44c06a0c24b11d58ece00" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.453834 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp"] Dec 10 14:58:38 crc kubenswrapper[4847]: E1210 14:58:38.454305 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d0cfc8-2361-4e74-bacc-379a98054d97" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.454332 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d0cfc8-2361-4e74-bacc-379a98054d97" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.454597 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d0cfc8-2361-4e74-bacc-379a98054d97" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.461908 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.464404 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.469003 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.469040 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.469060 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.469224 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.489064 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp"] Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.552999 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.553075 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.553178 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.553232 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.553314 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdchg\" (UniqueName: \"kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.654929 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdchg\" (UniqueName: \"kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.655036 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.655068 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.655126 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.655159 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.659486 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.659958 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.668886 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.669380 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.676154 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdchg\" (UniqueName: \"kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:38 crc kubenswrapper[4847]: I1210 14:58:38.787142 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 14:58:39 crc kubenswrapper[4847]: I1210 14:58:39.010464 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j9qsn" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="registry-server" probeResult="failure" output=< Dec 10 14:58:39 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 14:58:39 crc kubenswrapper[4847]: > Dec 10 14:58:39 crc kubenswrapper[4847]: I1210 14:58:39.378362 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp"] Dec 10 14:58:39 crc kubenswrapper[4847]: W1210 14:58:39.390399 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb5a7f57_0eb5_47f9_ae7c_173f849c7a02.slice/crio-3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90 WatchSource:0}: Error finding container 3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90: Status 404 returned error can't find the container with id 3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90 Dec 10 14:58:40 crc kubenswrapper[4847]: I1210 14:58:40.328990 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" event={"ID":"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02","Type":"ContainerStarted","Data":"3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90"} Dec 10 14:58:41 crc kubenswrapper[4847]: I1210 14:58:41.339356 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" event={"ID":"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02","Type":"ContainerStarted","Data":"c8f1c8833fde9802ef9cd366b1c68246f9d3030e3a16c9e611d9c0739416806e"} Dec 10 14:58:41 crc kubenswrapper[4847]: I1210 14:58:41.358501 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" podStartSLOduration=2.622212062 podStartE2EDuration="3.358480542s" podCreationTimestamp="2025-12-10 14:58:38 +0000 UTC" firstStartedPulling="2025-12-10 14:58:39.39461798 +0000 UTC m=+2068.963835610" lastFinishedPulling="2025-12-10 14:58:40.13088646 +0000 UTC m=+2069.700104090" observedRunningTime="2025-12-10 14:58:41.355813379 +0000 UTC m=+2070.925031019" watchObservedRunningTime="2025-12-10 14:58:41.358480542 +0000 UTC m=+2070.927698172" Dec 10 14:58:47 crc kubenswrapper[4847]: I1210 14:58:47.996036 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:48 crc kubenswrapper[4847]: I1210 14:58:48.055690 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:48 crc kubenswrapper[4847]: I1210 14:58:48.233390 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:49 crc kubenswrapper[4847]: I1210 14:58:49.423549 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j9qsn" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="registry-server" containerID="cri-o://71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f" gracePeriod=2 Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.206628 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.296892 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md84p\" (UniqueName: \"kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p\") pod \"b7a86ec2-573f-4b25-938b-3014dae2efb9\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.297082 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content\") pod \"b7a86ec2-573f-4b25-938b-3014dae2efb9\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.297166 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities\") pod \"b7a86ec2-573f-4b25-938b-3014dae2efb9\" (UID: \"b7a86ec2-573f-4b25-938b-3014dae2efb9\") " Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.298700 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities" (OuterVolumeSpecName: "utilities") pod "b7a86ec2-573f-4b25-938b-3014dae2efb9" (UID: "b7a86ec2-573f-4b25-938b-3014dae2efb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.311013 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p" (OuterVolumeSpecName: "kube-api-access-md84p") pod "b7a86ec2-573f-4b25-938b-3014dae2efb9" (UID: "b7a86ec2-573f-4b25-938b-3014dae2efb9"). InnerVolumeSpecName "kube-api-access-md84p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.399231 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.399274 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md84p\" (UniqueName: \"kubernetes.io/projected/b7a86ec2-573f-4b25-938b-3014dae2efb9-kube-api-access-md84p\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.437587 4847 generic.go:334] "Generic (PLEG): container finished" podID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerID="71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f" exitCode=0 Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.437645 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerDied","Data":"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f"} Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.437671 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9qsn" event={"ID":"b7a86ec2-573f-4b25-938b-3014dae2efb9","Type":"ContainerDied","Data":"05fc228d5a4ac6e5d6ff39ddd3fa5f02a3a127e304ddec042cdb734da3846105"} Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.437689 4847 scope.go:117] "RemoveContainer" containerID="71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.437858 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9qsn" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.463637 4847 scope.go:117] "RemoveContainer" containerID="e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.483746 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7a86ec2-573f-4b25-938b-3014dae2efb9" (UID: "b7a86ec2-573f-4b25-938b-3014dae2efb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.488678 4847 scope.go:117] "RemoveContainer" containerID="d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.501945 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7a86ec2-573f-4b25-938b-3014dae2efb9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.527792 4847 scope.go:117] "RemoveContainer" containerID="71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f" Dec 10 14:58:50 crc kubenswrapper[4847]: E1210 14:58:50.528414 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f\": container with ID starting with 71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f not found: ID does not exist" containerID="71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.528478 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f"} err="failed to get container status \"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f\": rpc error: code = NotFound desc = could not find container \"71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f\": container with ID starting with 71f255234b2bda8b10780981d051e3fa351e3d704abfafe8643349616662e92f not found: ID does not exist" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.528517 4847 scope.go:117] "RemoveContainer" containerID="e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de" Dec 10 14:58:50 crc kubenswrapper[4847]: E1210 14:58:50.529018 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de\": container with ID starting with e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de not found: ID does not exist" containerID="e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.529063 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de"} err="failed to get container status \"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de\": rpc error: code = NotFound desc = could not find container \"e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de\": container with ID starting with e1259ff5e1116cc260d9ab87acf42294170793641f8489b347fa043d601e43de not found: ID does not exist" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.529096 4847 scope.go:117] "RemoveContainer" containerID="d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0" Dec 10 14:58:50 crc kubenswrapper[4847]: E1210 14:58:50.529886 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0\": container with ID starting with d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0 not found: ID does not exist" containerID="d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.529922 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0"} err="failed to get container status \"d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0\": rpc error: code = NotFound desc = could not find container \"d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0\": container with ID starting with d1fd659b4be0e737ad96d0b1588e57672213849bf96308272da5e33ac20b49b0 not found: ID does not exist" Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.772519 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:50 crc kubenswrapper[4847]: I1210 14:58:50.779835 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j9qsn"] Dec 10 14:58:52 crc kubenswrapper[4847]: I1210 14:58:52.776129 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" path="/var/lib/kubelet/pods/b7a86ec2-573f-4b25-938b-3014dae2efb9/volumes" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.155208 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv"] Dec 10 15:00:00 crc kubenswrapper[4847]: E1210 15:00:00.156305 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="extract-content" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.156321 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="extract-content" Dec 10 15:00:00 crc kubenswrapper[4847]: E1210 15:00:00.156344 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="registry-server" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.156352 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="registry-server" Dec 10 15:00:00 crc kubenswrapper[4847]: E1210 15:00:00.156388 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="extract-utilities" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.156397 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="extract-utilities" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.156632 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a86ec2-573f-4b25-938b-3014dae2efb9" containerName="registry-server" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.157391 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.160901 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.160984 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.169077 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv"] Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.206379 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.206501 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drx7x\" (UniqueName: \"kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.206597 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.308093 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drx7x\" (UniqueName: \"kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.308202 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.308260 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.309101 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.317324 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.325320 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drx7x\" (UniqueName: \"kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x\") pod \"collect-profiles-29422980-vtxjv\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.483268 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:00 crc kubenswrapper[4847]: I1210 15:00:00.950975 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv"] Dec 10 15:00:01 crc kubenswrapper[4847]: I1210 15:00:01.011049 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:00:01 crc kubenswrapper[4847]: I1210 15:00:01.011640 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:00:01 crc kubenswrapper[4847]: I1210 15:00:01.412836 4847 generic.go:334] "Generic (PLEG): container finished" podID="8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" containerID="82aaca04c4c853195192392eceb53d70ba60690070bcd0f12bb851af04b15ece" exitCode=0 Dec 10 15:00:01 crc kubenswrapper[4847]: I1210 15:00:01.412892 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" event={"ID":"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7","Type":"ContainerDied","Data":"82aaca04c4c853195192392eceb53d70ba60690070bcd0f12bb851af04b15ece"} Dec 10 15:00:01 crc kubenswrapper[4847]: I1210 15:00:01.412922 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" event={"ID":"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7","Type":"ContainerStarted","Data":"1c57aa95beddb16eb5732726c806c3e8ee3e71707d178a8bbdce628e3e5e1dab"} Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.806103 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.976922 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume\") pod \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.977491 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drx7x\" (UniqueName: \"kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x\") pod \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.977598 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume\") pod \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\" (UID: \"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7\") " Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.978686 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" (UID: "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.986390 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" (UID: "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:00:02 crc kubenswrapper[4847]: I1210 15:00:02.992076 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x" (OuterVolumeSpecName: "kube-api-access-drx7x") pod "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" (UID: "8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7"). InnerVolumeSpecName "kube-api-access-drx7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.079783 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.079822 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.079834 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drx7x\" (UniqueName: \"kubernetes.io/projected/8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7-kube-api-access-drx7x\") on node \"crc\" DevicePath \"\"" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.430520 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" event={"ID":"8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7","Type":"ContainerDied","Data":"1c57aa95beddb16eb5732726c806c3e8ee3e71707d178a8bbdce628e3e5e1dab"} Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.430573 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c57aa95beddb16eb5732726c806c3e8ee3e71707d178a8bbdce628e3e5e1dab" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.430635 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422980-vtxjv" Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.882670 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk"] Dec 10 15:00:03 crc kubenswrapper[4847]: I1210 15:00:03.890921 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422935-68ckk"] Dec 10 15:00:04 crc kubenswrapper[4847]: I1210 15:00:04.772265 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69b03596-eace-45ce-ad89-bd52e7fd5766" path="/var/lib/kubelet/pods/69b03596-eace-45ce-ad89-bd52e7fd5766/volumes" Dec 10 15:00:24 crc kubenswrapper[4847]: I1210 15:00:24.229478 4847 scope.go:117] "RemoveContainer" containerID="5472ceb72d8aade48384de41aa34f010fe3b922e7059dce135fe308cd4d21621" Dec 10 15:00:31 crc kubenswrapper[4847]: I1210 15:00:31.011907 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:00:31 crc kubenswrapper[4847]: I1210 15:00:31.012595 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.160953 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29422981-q652q"] Dec 10 15:01:00 crc kubenswrapper[4847]: E1210 15:01:00.162206 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" containerName="collect-profiles" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.162226 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" containerName="collect-profiles" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.162460 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf38e1e-cbb6-4dd8-8682-a9b2e01c17f7" containerName="collect-profiles" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.163425 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.174924 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29422981-q652q"] Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.309865 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzq2r\" (UniqueName: \"kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.310107 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.310204 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.310419 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.412503 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.412599 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.412675 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.412784 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzq2r\" (UniqueName: \"kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.419611 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.420374 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.421495 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.436927 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzq2r\" (UniqueName: \"kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r\") pod \"keystone-cron-29422981-q652q\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.502107 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:00 crc kubenswrapper[4847]: I1210 15:01:00.992156 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29422981-q652q"] Dec 10 15:01:00 crc kubenswrapper[4847]: W1210 15:01:00.995879 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0790548_d5d8_4688_9120_bb942f395fd2.slice/crio-7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a WatchSource:0}: Error finding container 7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a: Status 404 returned error can't find the container with id 7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a Dec 10 15:01:01 crc kubenswrapper[4847]: I1210 15:01:01.011255 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:01:01 crc kubenswrapper[4847]: I1210 15:01:01.011326 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:01:01 crc kubenswrapper[4847]: I1210 15:01:01.011384 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:01:01 crc kubenswrapper[4847]: I1210 15:01:01.012359 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:01:01 crc kubenswrapper[4847]: I1210 15:01:01.012452 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" gracePeriod=600 Dec 10 15:01:01 crc kubenswrapper[4847]: E1210 15:01:01.170099 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.003291 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" exitCode=0 Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.003361 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d"} Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.003702 4847 scope.go:117] "RemoveContainer" containerID="f137860c78bac821a85f6a8ff099e7492e0c7726affeef91b90f61dae9c43c88" Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.004523 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:01:02 crc kubenswrapper[4847]: E1210 15:01:02.004990 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.009368 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422981-q652q" event={"ID":"b0790548-d5d8-4688-9120-bb942f395fd2","Type":"ContainerStarted","Data":"5065262b298daf8be08220a73b8b0c73f053fd8979175ff343f9de00f20f2c4e"} Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.009405 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422981-q652q" event={"ID":"b0790548-d5d8-4688-9120-bb942f395fd2","Type":"ContainerStarted","Data":"7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a"} Dec 10 15:01:02 crc kubenswrapper[4847]: I1210 15:01:02.057255 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29422981-q652q" podStartSLOduration=2.057235786 podStartE2EDuration="2.057235786s" podCreationTimestamp="2025-12-10 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 15:01:02.046826809 +0000 UTC m=+2211.616044439" watchObservedRunningTime="2025-12-10 15:01:02.057235786 +0000 UTC m=+2211.626453416" Dec 10 15:01:04 crc kubenswrapper[4847]: I1210 15:01:04.028532 4847 generic.go:334] "Generic (PLEG): container finished" podID="b0790548-d5d8-4688-9120-bb942f395fd2" containerID="5065262b298daf8be08220a73b8b0c73f053fd8979175ff343f9de00f20f2c4e" exitCode=0 Dec 10 15:01:04 crc kubenswrapper[4847]: I1210 15:01:04.028623 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422981-q652q" event={"ID":"b0790548-d5d8-4688-9120-bb942f395fd2","Type":"ContainerDied","Data":"5065262b298daf8be08220a73b8b0c73f053fd8979175ff343f9de00f20f2c4e"} Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.418574 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.564804 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys\") pod \"b0790548-d5d8-4688-9120-bb942f395fd2\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.565266 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzq2r\" (UniqueName: \"kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r\") pod \"b0790548-d5d8-4688-9120-bb942f395fd2\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.566178 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle\") pod \"b0790548-d5d8-4688-9120-bb942f395fd2\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.566855 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data\") pod \"b0790548-d5d8-4688-9120-bb942f395fd2\" (UID: \"b0790548-d5d8-4688-9120-bb942f395fd2\") " Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.571160 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r" (OuterVolumeSpecName: "kube-api-access-fzq2r") pod "b0790548-d5d8-4688-9120-bb942f395fd2" (UID: "b0790548-d5d8-4688-9120-bb942f395fd2"). InnerVolumeSpecName "kube-api-access-fzq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.572456 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b0790548-d5d8-4688-9120-bb942f395fd2" (UID: "b0790548-d5d8-4688-9120-bb942f395fd2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.602060 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0790548-d5d8-4688-9120-bb942f395fd2" (UID: "b0790548-d5d8-4688-9120-bb942f395fd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.626686 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data" (OuterVolumeSpecName: "config-data") pod "b0790548-d5d8-4688-9120-bb942f395fd2" (UID: "b0790548-d5d8-4688-9120-bb942f395fd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.669946 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzq2r\" (UniqueName: \"kubernetes.io/projected/b0790548-d5d8-4688-9120-bb942f395fd2-kube-api-access-fzq2r\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.670008 4847 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.670033 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:05 crc kubenswrapper[4847]: I1210 15:01:05.670051 4847 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b0790548-d5d8-4688-9120-bb942f395fd2-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:06 crc kubenswrapper[4847]: I1210 15:01:06.050406 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29422981-q652q" event={"ID":"b0790548-d5d8-4688-9120-bb942f395fd2","Type":"ContainerDied","Data":"7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a"} Dec 10 15:01:06 crc kubenswrapper[4847]: I1210 15:01:06.050450 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7649056d991f60f0a8a4e440158f66e902aa9625b4cd20b7851e5b4aebec416a" Dec 10 15:01:06 crc kubenswrapper[4847]: I1210 15:01:06.050526 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29422981-q652q" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.119151 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:15 crc kubenswrapper[4847]: E1210 15:01:15.121459 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0790548-d5d8-4688-9120-bb942f395fd2" containerName="keystone-cron" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.121480 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0790548-d5d8-4688-9120-bb942f395fd2" containerName="keystone-cron" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.121731 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0790548-d5d8-4688-9120-bb942f395fd2" containerName="keystone-cron" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.123025 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.150552 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.160291 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.160530 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzwv\" (UniqueName: \"kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.160578 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.261972 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.262080 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzwv\" (UniqueName: \"kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.262107 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.262734 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.262970 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.287511 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzwv\" (UniqueName: \"kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv\") pod \"community-operators-kz5pz\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.451903 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:15 crc kubenswrapper[4847]: I1210 15:01:15.759824 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:01:15 crc kubenswrapper[4847]: E1210 15:01:15.760462 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:01:16 crc kubenswrapper[4847]: I1210 15:01:16.003880 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:16 crc kubenswrapper[4847]: W1210 15:01:16.007336 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2dd8460_a0bb_424f_972e_dde0e36779d5.slice/crio-9f4ee4aee034d19699019b24a7e7ad87c6e8cfaa7237651d1a9b9b3629d3040c WatchSource:0}: Error finding container 9f4ee4aee034d19699019b24a7e7ad87c6e8cfaa7237651d1a9b9b3629d3040c: Status 404 returned error can't find the container with id 9f4ee4aee034d19699019b24a7e7ad87c6e8cfaa7237651d1a9b9b3629d3040c Dec 10 15:01:16 crc kubenswrapper[4847]: I1210 15:01:16.140771 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerStarted","Data":"9f4ee4aee034d19699019b24a7e7ad87c6e8cfaa7237651d1a9b9b3629d3040c"} Dec 10 15:01:17 crc kubenswrapper[4847]: I1210 15:01:17.156430 4847 generic.go:334] "Generic (PLEG): container finished" podID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerID="f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745" exitCode=0 Dec 10 15:01:17 crc kubenswrapper[4847]: I1210 15:01:17.156754 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerDied","Data":"f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745"} Dec 10 15:01:19 crc kubenswrapper[4847]: I1210 15:01:19.173452 4847 generic.go:334] "Generic (PLEG): container finished" podID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerID="870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80" exitCode=0 Dec 10 15:01:19 crc kubenswrapper[4847]: I1210 15:01:19.173564 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerDied","Data":"870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80"} Dec 10 15:01:21 crc kubenswrapper[4847]: I1210 15:01:21.193037 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerStarted","Data":"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656"} Dec 10 15:01:21 crc kubenswrapper[4847]: I1210 15:01:21.207395 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kz5pz" podStartSLOduration=3.200786281 podStartE2EDuration="6.207376434s" podCreationTimestamp="2025-12-10 15:01:15 +0000 UTC" firstStartedPulling="2025-12-10 15:01:17.158355714 +0000 UTC m=+2226.727573344" lastFinishedPulling="2025-12-10 15:01:20.164945867 +0000 UTC m=+2229.734163497" observedRunningTime="2025-12-10 15:01:21.206577783 +0000 UTC m=+2230.775795433" watchObservedRunningTime="2025-12-10 15:01:21.207376434 +0000 UTC m=+2230.776594064" Dec 10 15:01:25 crc kubenswrapper[4847]: I1210 15:01:25.452585 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:25 crc kubenswrapper[4847]: I1210 15:01:25.452942 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:25 crc kubenswrapper[4847]: I1210 15:01:25.503829 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:26 crc kubenswrapper[4847]: I1210 15:01:26.295171 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:26 crc kubenswrapper[4847]: I1210 15:01:26.349230 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:28 crc kubenswrapper[4847]: I1210 15:01:28.252509 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kz5pz" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="registry-server" containerID="cri-o://6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656" gracePeriod=2 Dec 10 15:01:28 crc kubenswrapper[4847]: E1210 15:01:28.475132 4847 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2dd8460_a0bb_424f_972e_dde0e36779d5.slice/crio-6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656.scope\": RecentStats: unable to find data in memory cache]" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.212780 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.263296 4847 generic.go:334] "Generic (PLEG): container finished" podID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerID="6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656" exitCode=0 Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.263345 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerDied","Data":"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656"} Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.263375 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kz5pz" event={"ID":"b2dd8460-a0bb-424f-972e-dde0e36779d5","Type":"ContainerDied","Data":"9f4ee4aee034d19699019b24a7e7ad87c6e8cfaa7237651d1a9b9b3629d3040c"} Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.263396 4847 scope.go:117] "RemoveContainer" containerID="6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.263394 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kz5pz" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.282432 4847 scope.go:117] "RemoveContainer" containerID="870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.307311 4847 scope.go:117] "RemoveContainer" containerID="f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.350615 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities\") pod \"b2dd8460-a0bb-424f-972e-dde0e36779d5\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.351551 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content\") pod \"b2dd8460-a0bb-424f-972e-dde0e36779d5\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.351706 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqzwv\" (UniqueName: \"kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv\") pod \"b2dd8460-a0bb-424f-972e-dde0e36779d5\" (UID: \"b2dd8460-a0bb-424f-972e-dde0e36779d5\") " Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.352173 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities" (OuterVolumeSpecName: "utilities") pod "b2dd8460-a0bb-424f-972e-dde0e36779d5" (UID: "b2dd8460-a0bb-424f-972e-dde0e36779d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.352584 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.352610 4847 scope.go:117] "RemoveContainer" containerID="6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656" Dec 10 15:01:29 crc kubenswrapper[4847]: E1210 15:01:29.353413 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656\": container with ID starting with 6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656 not found: ID does not exist" containerID="6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.353450 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656"} err="failed to get container status \"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656\": rpc error: code = NotFound desc = could not find container \"6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656\": container with ID starting with 6a3abe88e85218f05bf0bb298f129f2358c828c9cc80ae0df99dfad3ff7ae656 not found: ID does not exist" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.353472 4847 scope.go:117] "RemoveContainer" containerID="870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80" Dec 10 15:01:29 crc kubenswrapper[4847]: E1210 15:01:29.354123 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80\": container with ID starting with 870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80 not found: ID does not exist" containerID="870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.354181 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80"} err="failed to get container status \"870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80\": rpc error: code = NotFound desc = could not find container \"870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80\": container with ID starting with 870c7e03c0ebbea05101ed8935e15748f515ad57b15fd4a81c33190596dc6b80 not found: ID does not exist" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.354218 4847 scope.go:117] "RemoveContainer" containerID="f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745" Dec 10 15:01:29 crc kubenswrapper[4847]: E1210 15:01:29.355294 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745\": container with ID starting with f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745 not found: ID does not exist" containerID="f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.355330 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745"} err="failed to get container status \"f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745\": rpc error: code = NotFound desc = could not find container \"f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745\": container with ID starting with f8f2d31848f3c7764833d1c675ed5bb85c2fda6bdf70d8996ce2676e0ed2f745 not found: ID does not exist" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.358469 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv" (OuterVolumeSpecName: "kube-api-access-zqzwv") pod "b2dd8460-a0bb-424f-972e-dde0e36779d5" (UID: "b2dd8460-a0bb-424f-972e-dde0e36779d5"). InnerVolumeSpecName "kube-api-access-zqzwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.403802 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2dd8460-a0bb-424f-972e-dde0e36779d5" (UID: "b2dd8460-a0bb-424f-972e-dde0e36779d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.454726 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2dd8460-a0bb-424f-972e-dde0e36779d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.454964 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqzwv\" (UniqueName: \"kubernetes.io/projected/b2dd8460-a0bb-424f-972e-dde0e36779d5-kube-api-access-zqzwv\") on node \"crc\" DevicePath \"\"" Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.598515 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.608183 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kz5pz"] Dec 10 15:01:29 crc kubenswrapper[4847]: I1210 15:01:29.761433 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:01:29 crc kubenswrapper[4847]: E1210 15:01:29.766311 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:01:30 crc kubenswrapper[4847]: I1210 15:01:30.772725 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" path="/var/lib/kubelet/pods/b2dd8460-a0bb-424f-972e-dde0e36779d5/volumes" Dec 10 15:01:40 crc kubenswrapper[4847]: I1210 15:01:40.766352 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:01:40 crc kubenswrapper[4847]: E1210 15:01:40.767338 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:01:53 crc kubenswrapper[4847]: I1210 15:01:53.761510 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:01:53 crc kubenswrapper[4847]: E1210 15:01:53.762926 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.572862 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kqks9"] Dec 10 15:02:06 crc kubenswrapper[4847]: E1210 15:02:06.573919 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="extract-content" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.573938 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="extract-content" Dec 10 15:02:06 crc kubenswrapper[4847]: E1210 15:02:06.573971 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="registry-server" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.573979 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="registry-server" Dec 10 15:02:06 crc kubenswrapper[4847]: E1210 15:02:06.573993 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="extract-utilities" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.574002 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="extract-utilities" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.574232 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2dd8460-a0bb-424f-972e-dde0e36779d5" containerName="registry-server" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.576075 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.603125 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqks9"] Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.746389 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-utilities\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.746466 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8k4l\" (UniqueName: \"kubernetes.io/projected/2f6069b4-37e0-452f-9361-f5da5a9a0f52-kube-api-access-d8k4l\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.746519 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-catalog-content\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.848194 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-catalog-content\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.848454 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-utilities\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.848494 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8k4l\" (UniqueName: \"kubernetes.io/projected/2f6069b4-37e0-452f-9361-f5da5a9a0f52-kube-api-access-d8k4l\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.848668 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-catalog-content\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.848874 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f6069b4-37e0-452f-9361-f5da5a9a0f52-utilities\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.872948 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8k4l\" (UniqueName: \"kubernetes.io/projected/2f6069b4-37e0-452f-9361-f5da5a9a0f52-kube-api-access-d8k4l\") pod \"certified-operators-kqks9\" (UID: \"2f6069b4-37e0-452f-9361-f5da5a9a0f52\") " pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:06 crc kubenswrapper[4847]: I1210 15:02:06.901805 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:07 crc kubenswrapper[4847]: I1210 15:02:07.434621 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqks9"] Dec 10 15:02:07 crc kubenswrapper[4847]: W1210 15:02:07.442438 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f6069b4_37e0_452f_9361_f5da5a9a0f52.slice/crio-dc24e1311d7257106cdd9322f2ac1e6e58f4fd466d98442575fcc1f2df12331d WatchSource:0}: Error finding container dc24e1311d7257106cdd9322f2ac1e6e58f4fd466d98442575fcc1f2df12331d: Status 404 returned error can't find the container with id dc24e1311d7257106cdd9322f2ac1e6e58f4fd466d98442575fcc1f2df12331d Dec 10 15:02:07 crc kubenswrapper[4847]: I1210 15:02:07.602152 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqks9" event={"ID":"2f6069b4-37e0-452f-9361-f5da5a9a0f52","Type":"ContainerStarted","Data":"dc24e1311d7257106cdd9322f2ac1e6e58f4fd466d98442575fcc1f2df12331d"} Dec 10 15:02:07 crc kubenswrapper[4847]: I1210 15:02:07.759837 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:02:07 crc kubenswrapper[4847]: E1210 15:02:07.760206 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:08 crc kubenswrapper[4847]: I1210 15:02:08.614151 4847 generic.go:334] "Generic (PLEG): container finished" podID="2f6069b4-37e0-452f-9361-f5da5a9a0f52" containerID="cfedae9557f976c314bda515280c5fa95fb6d6a7380122be770ddeae0c3f943d" exitCode=0 Dec 10 15:02:08 crc kubenswrapper[4847]: I1210 15:02:08.614269 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqks9" event={"ID":"2f6069b4-37e0-452f-9361-f5da5a9a0f52","Type":"ContainerDied","Data":"cfedae9557f976c314bda515280c5fa95fb6d6a7380122be770ddeae0c3f943d"} Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.372964 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.375310 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.385840 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.519042 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.519112 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.519174 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8kbg\" (UniqueName: \"kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.620960 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.621361 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.621416 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8kbg\" (UniqueName: \"kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.621557 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.621882 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.646794 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8kbg\" (UniqueName: \"kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg\") pod \"redhat-marketplace-7fdv2\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:09 crc kubenswrapper[4847]: I1210 15:02:09.701941 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:10 crc kubenswrapper[4847]: I1210 15:02:10.189698 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:10 crc kubenswrapper[4847]: W1210 15:02:10.198277 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b4ae0d4_5659_4f3c_9907_c1cd0e97ab34.slice/crio-d05b55b112d0bac749319a96b03fc4ae895e0025fdc53061506a1ebeac423b43 WatchSource:0}: Error finding container d05b55b112d0bac749319a96b03fc4ae895e0025fdc53061506a1ebeac423b43: Status 404 returned error can't find the container with id d05b55b112d0bac749319a96b03fc4ae895e0025fdc53061506a1ebeac423b43 Dec 10 15:02:10 crc kubenswrapper[4847]: I1210 15:02:10.636967 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerID="e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43" exitCode=0 Dec 10 15:02:10 crc kubenswrapper[4847]: I1210 15:02:10.637177 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerDied","Data":"e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43"} Dec 10 15:02:10 crc kubenswrapper[4847]: I1210 15:02:10.637345 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerStarted","Data":"d05b55b112d0bac749319a96b03fc4ae895e0025fdc53061506a1ebeac423b43"} Dec 10 15:02:11 crc kubenswrapper[4847]: I1210 15:02:11.650563 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerStarted","Data":"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6"} Dec 10 15:02:12 crc kubenswrapper[4847]: I1210 15:02:12.661073 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerID="b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6" exitCode=0 Dec 10 15:02:12 crc kubenswrapper[4847]: I1210 15:02:12.661190 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerDied","Data":"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6"} Dec 10 15:02:13 crc kubenswrapper[4847]: I1210 15:02:13.674763 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqks9" event={"ID":"2f6069b4-37e0-452f-9361-f5da5a9a0f52","Type":"ContainerStarted","Data":"d0c9feaacc915a2f370a7e1d35b94f6a188c0d90ff2f4f6f9ed200a11ac45127"} Dec 10 15:02:14 crc kubenswrapper[4847]: I1210 15:02:14.686502 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerStarted","Data":"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77"} Dec 10 15:02:14 crc kubenswrapper[4847]: I1210 15:02:14.688737 4847 generic.go:334] "Generic (PLEG): container finished" podID="2f6069b4-37e0-452f-9361-f5da5a9a0f52" containerID="d0c9feaacc915a2f370a7e1d35b94f6a188c0d90ff2f4f6f9ed200a11ac45127" exitCode=0 Dec 10 15:02:14 crc kubenswrapper[4847]: I1210 15:02:14.688765 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqks9" event={"ID":"2f6069b4-37e0-452f-9361-f5da5a9a0f52","Type":"ContainerDied","Data":"d0c9feaacc915a2f370a7e1d35b94f6a188c0d90ff2f4f6f9ed200a11ac45127"} Dec 10 15:02:14 crc kubenswrapper[4847]: I1210 15:02:14.713175 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7fdv2" podStartSLOduration=2.669493117 podStartE2EDuration="5.713148816s" podCreationTimestamp="2025-12-10 15:02:09 +0000 UTC" firstStartedPulling="2025-12-10 15:02:10.639215182 +0000 UTC m=+2280.208432812" lastFinishedPulling="2025-12-10 15:02:13.682870881 +0000 UTC m=+2283.252088511" observedRunningTime="2025-12-10 15:02:14.701328332 +0000 UTC m=+2284.270545982" watchObservedRunningTime="2025-12-10 15:02:14.713148816 +0000 UTC m=+2284.282366446" Dec 10 15:02:15 crc kubenswrapper[4847]: I1210 15:02:15.700153 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kqks9" event={"ID":"2f6069b4-37e0-452f-9361-f5da5a9a0f52","Type":"ContainerStarted","Data":"95307abd3afcc888d73d37be7e54c9e9ab3946cd3370441ef47d9cdd51e23e7e"} Dec 10 15:02:15 crc kubenswrapper[4847]: I1210 15:02:15.722045 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kqks9" podStartSLOduration=3.017056225 podStartE2EDuration="9.722025771s" podCreationTimestamp="2025-12-10 15:02:06 +0000 UTC" firstStartedPulling="2025-12-10 15:02:08.616790258 +0000 UTC m=+2278.186007888" lastFinishedPulling="2025-12-10 15:02:15.321759804 +0000 UTC m=+2284.890977434" observedRunningTime="2025-12-10 15:02:15.717357196 +0000 UTC m=+2285.286574846" watchObservedRunningTime="2025-12-10 15:02:15.722025771 +0000 UTC m=+2285.291243401" Dec 10 15:02:16 crc kubenswrapper[4847]: I1210 15:02:16.903498 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:16 crc kubenswrapper[4847]: I1210 15:02:16.903804 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:17 crc kubenswrapper[4847]: I1210 15:02:17.957926 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kqks9" podUID="2f6069b4-37e0-452f-9361-f5da5a9a0f52" containerName="registry-server" probeResult="failure" output=< Dec 10 15:02:17 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 15:02:17 crc kubenswrapper[4847]: > Dec 10 15:02:19 crc kubenswrapper[4847]: I1210 15:02:19.702264 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:19 crc kubenswrapper[4847]: I1210 15:02:19.702325 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:19 crc kubenswrapper[4847]: I1210 15:02:19.757591 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:19 crc kubenswrapper[4847]: I1210 15:02:19.759632 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:02:19 crc kubenswrapper[4847]: E1210 15:02:19.759929 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:19 crc kubenswrapper[4847]: I1210 15:02:19.807781 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:20 crc kubenswrapper[4847]: I1210 15:02:20.001568 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:21 crc kubenswrapper[4847]: I1210 15:02:21.760118 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7fdv2" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="registry-server" containerID="cri-o://2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77" gracePeriod=2 Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.223968 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.286107 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content\") pod \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.286167 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities\") pod \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.286192 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8kbg\" (UniqueName: \"kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg\") pod \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\" (UID: \"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34\") " Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.287055 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities" (OuterVolumeSpecName: "utilities") pod "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" (UID: "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.292965 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg" (OuterVolumeSpecName: "kube-api-access-m8kbg") pod "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" (UID: "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34"). InnerVolumeSpecName "kube-api-access-m8kbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.309125 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" (UID: "6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.387451 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.387484 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.387493 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8kbg\" (UniqueName: \"kubernetes.io/projected/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34-kube-api-access-m8kbg\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.771826 4847 generic.go:334] "Generic (PLEG): container finished" podID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerID="2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77" exitCode=0 Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.771889 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fdv2" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.771875 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerDied","Data":"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77"} Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.771959 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fdv2" event={"ID":"6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34","Type":"ContainerDied","Data":"d05b55b112d0bac749319a96b03fc4ae895e0025fdc53061506a1ebeac423b43"} Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.771983 4847 scope.go:117] "RemoveContainer" containerID="2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.813656 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.814485 4847 scope.go:117] "RemoveContainer" containerID="b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.825477 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fdv2"] Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.844409 4847 scope.go:117] "RemoveContainer" containerID="e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.881845 4847 scope.go:117] "RemoveContainer" containerID="2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77" Dec 10 15:02:22 crc kubenswrapper[4847]: E1210 15:02:22.882255 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77\": container with ID starting with 2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77 not found: ID does not exist" containerID="2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.882291 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77"} err="failed to get container status \"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77\": rpc error: code = NotFound desc = could not find container \"2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77\": container with ID starting with 2da258a1c0767e03113a5ab59e3f88e45fa773dfabf20a42f463b6e1c851ec77 not found: ID does not exist" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.882315 4847 scope.go:117] "RemoveContainer" containerID="b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6" Dec 10 15:02:22 crc kubenswrapper[4847]: E1210 15:02:22.882670 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6\": container with ID starting with b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6 not found: ID does not exist" containerID="b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.882695 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6"} err="failed to get container status \"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6\": rpc error: code = NotFound desc = could not find container \"b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6\": container with ID starting with b95f00cc6817657389ae89df8b858cfe801439b69ff214fd18f027158bc985a6 not found: ID does not exist" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.882732 4847 scope.go:117] "RemoveContainer" containerID="e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43" Dec 10 15:02:22 crc kubenswrapper[4847]: E1210 15:02:22.882997 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43\": container with ID starting with e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43 not found: ID does not exist" containerID="e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43" Dec 10 15:02:22 crc kubenswrapper[4847]: I1210 15:02:22.883028 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43"} err="failed to get container status \"e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43\": rpc error: code = NotFound desc = could not find container \"e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43\": container with ID starting with e3b257367758e27a2f3aae456e8062813884693415de874f2e0441a85aa04f43 not found: ID does not exist" Dec 10 15:02:24 crc kubenswrapper[4847]: I1210 15:02:24.771219 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" path="/var/lib/kubelet/pods/6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34/volumes" Dec 10 15:02:26 crc kubenswrapper[4847]: I1210 15:02:26.949566 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:26.999952 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kqks9" Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:27.066144 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kqks9"] Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:27.193081 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:27.193399 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s6wpx" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="registry-server" containerID="cri-o://2fea80235bef64c4646a08eb3f390fe08efbd938575e71c1d2b19878cf27f51d" gracePeriod=2 Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:27.814122 4847 generic.go:334] "Generic (PLEG): container finished" podID="0d91abd9-2ff0-4696-8942-32149e698707" containerID="2fea80235bef64c4646a08eb3f390fe08efbd938575e71c1d2b19878cf27f51d" exitCode=0 Dec 10 15:02:27 crc kubenswrapper[4847]: I1210 15:02:27.814306 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerDied","Data":"2fea80235bef64c4646a08eb3f390fe08efbd938575e71c1d2b19878cf27f51d"} Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.210148 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.302609 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content\") pod \"0d91abd9-2ff0-4696-8942-32149e698707\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.302837 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities\") pod \"0d91abd9-2ff0-4696-8942-32149e698707\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.302969 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqfch\" (UniqueName: \"kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch\") pod \"0d91abd9-2ff0-4696-8942-32149e698707\" (UID: \"0d91abd9-2ff0-4696-8942-32149e698707\") " Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.303498 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities" (OuterVolumeSpecName: "utilities") pod "0d91abd9-2ff0-4696-8942-32149e698707" (UID: "0d91abd9-2ff0-4696-8942-32149e698707"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.314905 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch" (OuterVolumeSpecName: "kube-api-access-fqfch") pod "0d91abd9-2ff0-4696-8942-32149e698707" (UID: "0d91abd9-2ff0-4696-8942-32149e698707"). InnerVolumeSpecName "kube-api-access-fqfch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.356162 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d91abd9-2ff0-4696-8942-32149e698707" (UID: "0d91abd9-2ff0-4696-8942-32149e698707"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.405268 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.405316 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqfch\" (UniqueName: \"kubernetes.io/projected/0d91abd9-2ff0-4696-8942-32149e698707-kube-api-access-fqfch\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.405329 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d91abd9-2ff0-4696-8942-32149e698707-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.840110 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s6wpx" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.840107 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s6wpx" event={"ID":"0d91abd9-2ff0-4696-8942-32149e698707","Type":"ContainerDied","Data":"329a5253d6d2842d980cf9f492fb5830fcb78425d9f9e71bfa6b9f0801ec2dab"} Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.841372 4847 scope.go:117] "RemoveContainer" containerID="2fea80235bef64c4646a08eb3f390fe08efbd938575e71c1d2b19878cf27f51d" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.881244 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.883059 4847 scope.go:117] "RemoveContainer" containerID="87f1d6aea560eba1ef453be09d695c66b811d04d13f70693cfa32579bf0926cf" Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.889238 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s6wpx"] Dec 10 15:02:28 crc kubenswrapper[4847]: I1210 15:02:28.906912 4847 scope.go:117] "RemoveContainer" containerID="366c087f642193fedbe6897b048b89acb3983aec4bccab1217a9a7156dde0718" Dec 10 15:02:30 crc kubenswrapper[4847]: I1210 15:02:30.771881 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d91abd9-2ff0-4696-8942-32149e698707" path="/var/lib/kubelet/pods/0d91abd9-2ff0-4696-8942-32149e698707/volumes" Dec 10 15:02:32 crc kubenswrapper[4847]: I1210 15:02:32.760602 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:02:32 crc kubenswrapper[4847]: E1210 15:02:32.761738 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:43 crc kubenswrapper[4847]: I1210 15:02:43.760767 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:02:43 crc kubenswrapper[4847]: E1210 15:02:43.761683 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:53 crc kubenswrapper[4847]: I1210 15:02:53.093351 4847 generic.go:334] "Generic (PLEG): container finished" podID="eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" containerID="c8f1c8833fde9802ef9cd366b1c68246f9d3030e3a16c9e611d9c0739416806e" exitCode=0 Dec 10 15:02:53 crc kubenswrapper[4847]: I1210 15:02:53.093468 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" event={"ID":"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02","Type":"ContainerDied","Data":"c8f1c8833fde9802ef9cd366b1c68246f9d3030e3a16c9e611d9c0739416806e"} Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.554316 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.625489 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdchg\" (UniqueName: \"kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg\") pod \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.625563 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0\") pod \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.625621 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory\") pod \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.625638 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key\") pod \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.625727 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle\") pod \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\" (UID: \"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02\") " Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.631983 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg" (OuterVolumeSpecName: "kube-api-access-qdchg") pod "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" (UID: "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02"). InnerVolumeSpecName "kube-api-access-qdchg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.633448 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" (UID: "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.657290 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" (UID: "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.662965 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" (UID: "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.670199 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory" (OuterVolumeSpecName: "inventory") pod "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" (UID: "eb5a7f57-0eb5-47f9-ae7c-173f849c7a02"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.729433 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdchg\" (UniqueName: \"kubernetes.io/projected/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-kube-api-access-qdchg\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.729470 4847 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.729480 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.729492 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:54 crc kubenswrapper[4847]: I1210 15:02:54.729502 4847 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb5a7f57-0eb5-47f9-ae7c-173f849c7a02-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.118599 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" event={"ID":"eb5a7f57-0eb5-47f9-ae7c-173f849c7a02","Type":"ContainerDied","Data":"3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90"} Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.118898 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f2be2fb8ee93ca1181792c418ee1749fc1480a8af0b059d53d25627ae555c90" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.118958 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.208959 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn"] Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209351 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="extract-content" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209367 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="extract-content" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209394 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="extract-content" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209403 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="extract-content" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209416 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="extract-utilities" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209423 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="extract-utilities" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209438 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209445 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209470 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209476 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209491 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="extract-utilities" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209499 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="extract-utilities" Dec 10 15:02:55 crc kubenswrapper[4847]: E1210 15:02:55.209513 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209521 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209735 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5a7f57-0eb5-47f9-ae7c-173f849c7a02" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209760 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d91abd9-2ff0-4696-8942-32149e698707" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.209772 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b4ae0d4-5659-4f3c-9907-c1cd0e97ab34" containerName="registry-server" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.210360 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213079 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213247 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213358 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213478 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213669 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213812 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.213918 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.224430 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn"] Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.237838 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.237880 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.237925 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.237955 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss965\" (UniqueName: \"kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.237980 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.238007 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.238032 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.238053 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.238081 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339551 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339622 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss965\" (UniqueName: \"kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339656 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339688 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339709 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339748 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339778 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339843 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.339866 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.340923 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.344364 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.344443 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.344862 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.344902 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.345158 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.345617 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.348346 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.357015 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss965\" (UniqueName: \"kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ct6mn\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.561953 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.890787 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn"] Dec 10 15:02:55 crc kubenswrapper[4847]: I1210 15:02:55.898444 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 15:02:56 crc kubenswrapper[4847]: I1210 15:02:56.127182 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" event={"ID":"9b520035-682d-43ae-acba-403455d78c9f","Type":"ContainerStarted","Data":"ecd9de6cf02dc91795e32a5357118f21a4b4a7ca2c97de9ccd651bf97e0db9c1"} Dec 10 15:02:56 crc kubenswrapper[4847]: I1210 15:02:56.761683 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:02:56 crc kubenswrapper[4847]: E1210 15:02:56.761992 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:02:57 crc kubenswrapper[4847]: I1210 15:02:57.140704 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" event={"ID":"9b520035-682d-43ae-acba-403455d78c9f","Type":"ContainerStarted","Data":"f09ae834d755d01f53cf53f7802cf40535588d06303efa2eff4f5f74ae4a3c73"} Dec 10 15:02:57 crc kubenswrapper[4847]: I1210 15:02:57.156928 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" podStartSLOduration=1.394220507 podStartE2EDuration="2.156905304s" podCreationTimestamp="2025-12-10 15:02:55 +0000 UTC" firstStartedPulling="2025-12-10 15:02:55.898180002 +0000 UTC m=+2325.467397632" lastFinishedPulling="2025-12-10 15:02:56.660864799 +0000 UTC m=+2326.230082429" observedRunningTime="2025-12-10 15:02:57.156255315 +0000 UTC m=+2326.725472945" watchObservedRunningTime="2025-12-10 15:02:57.156905304 +0000 UTC m=+2326.726122934" Dec 10 15:03:07 crc kubenswrapper[4847]: I1210 15:03:07.760709 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:03:07 crc kubenswrapper[4847]: E1210 15:03:07.761863 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:03:21 crc kubenswrapper[4847]: I1210 15:03:21.759650 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:03:21 crc kubenswrapper[4847]: E1210 15:03:21.760469 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:03:36 crc kubenswrapper[4847]: I1210 15:03:36.760324 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:03:36 crc kubenswrapper[4847]: E1210 15:03:36.761301 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:03:47 crc kubenswrapper[4847]: I1210 15:03:47.760386 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:03:47 crc kubenswrapper[4847]: E1210 15:03:47.761654 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:04:02 crc kubenswrapper[4847]: I1210 15:04:02.760678 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:04:02 crc kubenswrapper[4847]: E1210 15:04:02.761682 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:04:14 crc kubenswrapper[4847]: I1210 15:04:14.759893 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:04:14 crc kubenswrapper[4847]: E1210 15:04:14.760901 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:04:29 crc kubenswrapper[4847]: I1210 15:04:29.760547 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:04:29 crc kubenswrapper[4847]: E1210 15:04:29.761512 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:04:44 crc kubenswrapper[4847]: I1210 15:04:44.760147 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:04:44 crc kubenswrapper[4847]: E1210 15:04:44.762631 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:04:56 crc kubenswrapper[4847]: I1210 15:04:56.759502 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:04:56 crc kubenswrapper[4847]: E1210 15:04:56.760470 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:05:08 crc kubenswrapper[4847]: I1210 15:05:08.759802 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:05:08 crc kubenswrapper[4847]: E1210 15:05:08.760536 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:05:21 crc kubenswrapper[4847]: I1210 15:05:21.759855 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:05:21 crc kubenswrapper[4847]: E1210 15:05:21.760803 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:05:35 crc kubenswrapper[4847]: I1210 15:05:35.589161 4847 generic.go:334] "Generic (PLEG): container finished" podID="9b520035-682d-43ae-acba-403455d78c9f" containerID="f09ae834d755d01f53cf53f7802cf40535588d06303efa2eff4f5f74ae4a3c73" exitCode=0 Dec 10 15:05:35 crc kubenswrapper[4847]: I1210 15:05:35.589250 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" event={"ID":"9b520035-682d-43ae-acba-403455d78c9f","Type":"ContainerDied","Data":"f09ae834d755d01f53cf53f7802cf40535588d06303efa2eff4f5f74ae4a3c73"} Dec 10 15:05:36 crc kubenswrapper[4847]: I1210 15:05:36.759494 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:05:36 crc kubenswrapper[4847]: E1210 15:05:36.760087 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.011317 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.136728 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss965\" (UniqueName: \"kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.136830 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.136856 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.136982 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.137019 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.137078 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.137959 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.138007 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.138032 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0\") pod \"9b520035-682d-43ae-acba-403455d78c9f\" (UID: \"9b520035-682d-43ae-acba-403455d78c9f\") " Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.155243 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.162629 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965" (OuterVolumeSpecName: "kube-api-access-ss965") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "kube-api-access-ss965". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.169280 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.171914 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory" (OuterVolumeSpecName: "inventory") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.172147 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.173101 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.177035 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.180130 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.190436 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9b520035-682d-43ae-acba-403455d78c9f" (UID: "9b520035-682d-43ae-acba-403455d78c9f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241151 4847 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241208 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241389 4847 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241415 4847 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9b520035-682d-43ae-acba-403455d78c9f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241431 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss965\" (UniqueName: \"kubernetes.io/projected/9b520035-682d-43ae-acba-403455d78c9f-kube-api-access-ss965\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241446 4847 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241458 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241470 4847 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.241482 4847 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9b520035-682d-43ae-acba-403455d78c9f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.611332 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" event={"ID":"9b520035-682d-43ae-acba-403455d78c9f","Type":"ContainerDied","Data":"ecd9de6cf02dc91795e32a5357118f21a4b4a7ca2c97de9ccd651bf97e0db9c1"} Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.611384 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecd9de6cf02dc91795e32a5357118f21a4b4a7ca2c97de9ccd651bf97e0db9c1" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.611469 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ct6mn" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.714276 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp"] Dec 10 15:05:37 crc kubenswrapper[4847]: E1210 15:05:37.714787 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b520035-682d-43ae-acba-403455d78c9f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.714818 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b520035-682d-43ae-acba-403455d78c9f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.715145 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b520035-682d-43ae-acba-403455d78c9f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.716329 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.720783 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.721032 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.721044 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.721443 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.722383 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s8cqg" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.724912 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp"] Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.855182 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856020 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl9dj\" (UniqueName: \"kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856082 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856179 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856445 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856536 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.856652 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.958365 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.958475 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.958595 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.958639 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.959971 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl9dj\" (UniqueName: \"kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.960038 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.960117 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.963188 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.963448 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.963513 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.972925 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.981827 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.984224 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:37 crc kubenswrapper[4847]: I1210 15:05:37.987518 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl9dj\" (UniqueName: \"kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:38 crc kubenswrapper[4847]: I1210 15:05:38.043130 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:05:38 crc kubenswrapper[4847]: I1210 15:05:38.539267 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp"] Dec 10 15:05:38 crc kubenswrapper[4847]: I1210 15:05:38.619454 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" event={"ID":"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55","Type":"ContainerStarted","Data":"970b36ecf4ddec0d04c601343a4703e549e8c05ab7471973a2c31d688e7b8d0f"} Dec 10 15:05:39 crc kubenswrapper[4847]: I1210 15:05:39.627481 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" event={"ID":"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55","Type":"ContainerStarted","Data":"b9a2440bbe315c3e469664e121bf099b8144a8c13f2a06551b4ac4f94f4daebb"} Dec 10 15:05:39 crc kubenswrapper[4847]: I1210 15:05:39.642167 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" podStartSLOduration=1.971934419 podStartE2EDuration="2.642142543s" podCreationTimestamp="2025-12-10 15:05:37 +0000 UTC" firstStartedPulling="2025-12-10 15:05:38.543400399 +0000 UTC m=+2488.112618019" lastFinishedPulling="2025-12-10 15:05:39.213608503 +0000 UTC m=+2488.782826143" observedRunningTime="2025-12-10 15:05:39.640587701 +0000 UTC m=+2489.209805341" watchObservedRunningTime="2025-12-10 15:05:39.642142543 +0000 UTC m=+2489.211360173" Dec 10 15:05:51 crc kubenswrapper[4847]: I1210 15:05:51.759954 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:05:51 crc kubenswrapper[4847]: E1210 15:05:51.762066 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:06:04 crc kubenswrapper[4847]: I1210 15:06:04.760768 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:06:05 crc kubenswrapper[4847]: I1210 15:06:05.882292 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a"} Dec 10 15:07:55 crc kubenswrapper[4847]: I1210 15:07:55.084500 4847 generic.go:334] "Generic (PLEG): container finished" podID="f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" containerID="b9a2440bbe315c3e469664e121bf099b8144a8c13f2a06551b4ac4f94f4daebb" exitCode=0 Dec 10 15:07:55 crc kubenswrapper[4847]: I1210 15:07:55.084641 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" event={"ID":"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55","Type":"ContainerDied","Data":"b9a2440bbe315c3e469664e121bf099b8144a8c13f2a06551b4ac4f94f4daebb"} Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.574040 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.683911 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.683999 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.684058 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.684226 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.684302 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.684595 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.684860 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl9dj\" (UniqueName: \"kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj\") pod \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\" (UID: \"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55\") " Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.697000 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj" (OuterVolumeSpecName: "kube-api-access-bl9dj") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "kube-api-access-bl9dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.700932 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.718842 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.721525 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory" (OuterVolumeSpecName: "inventory") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.730023 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.734210 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.743621 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" (UID: "f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787568 4847 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787603 4847 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787614 4847 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787625 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl9dj\" (UniqueName: \"kubernetes.io/projected/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-kube-api-access-bl9dj\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787635 4847 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787644 4847 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:56 crc kubenswrapper[4847]: I1210 15:07:56.787654 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 15:07:57 crc kubenswrapper[4847]: I1210 15:07:57.108144 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" event={"ID":"f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55","Type":"ContainerDied","Data":"970b36ecf4ddec0d04c601343a4703e549e8c05ab7471973a2c31d688e7b8d0f"} Dec 10 15:07:57 crc kubenswrapper[4847]: I1210 15:07:57.108483 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970b36ecf4ddec0d04c601343a4703e549e8c05ab7471973a2c31d688e7b8d0f" Dec 10 15:07:57 crc kubenswrapper[4847]: I1210 15:07:57.108221 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp" Dec 10 15:08:31 crc kubenswrapper[4847]: I1210 15:08:31.011315 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:08:31 crc kubenswrapper[4847]: I1210 15:08:31.012902 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.211343 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:08:50 crc kubenswrapper[4847]: E1210 15:08:50.213189 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.213209 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.213450 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.215506 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.229170 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.345880 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.345991 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.346027 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khfsx\" (UniqueName: \"kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.447789 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.447881 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.447907 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khfsx\" (UniqueName: \"kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.448384 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.448398 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.467734 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khfsx\" (UniqueName: \"kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx\") pod \"redhat-operators-5t9xh\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:50 crc kubenswrapper[4847]: I1210 15:08:50.553314 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.031391 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.336011 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.337523 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.341772 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.341961 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.342059 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-ncmp9" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.342575 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.375028 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497127 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497176 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497220 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497571 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497684 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497782 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497815 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497893 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9thp\" (UniqueName: \"kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.497954 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600106 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600148 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600174 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600191 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600225 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9thp\" (UniqueName: \"kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600243 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600283 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600303 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600339 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600682 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600764 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.600918 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.601508 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.602058 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.606025 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.606485 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.607233 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.619668 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9thp\" (UniqueName: \"kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.642413 4847 generic.go:334] "Generic (PLEG): container finished" podID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerID="c01eb94a0f840cd8e465de6c351e25cdde5d343d5b6666253681a65afade8a5a" exitCode=0 Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.642457 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerDied","Data":"c01eb94a0f840cd8e465de6c351e25cdde5d343d5b6666253681a65afade8a5a"} Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.642484 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerStarted","Data":"af30346f6cd7f8b9ae192b32fe64e745cc853bbe36b9da6be63a86fcec3ea636"} Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.643745 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " pod="openstack/tempest-tests-tempest" Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.644547 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 15:08:51 crc kubenswrapper[4847]: I1210 15:08:51.675911 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 15:08:52 crc kubenswrapper[4847]: I1210 15:08:52.125582 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 15:08:52 crc kubenswrapper[4847]: W1210 15:08:52.144810 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1240adcf_343f_4c3e_8323_d4ca80e2b99d.slice/crio-9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29 WatchSource:0}: Error finding container 9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29: Status 404 returned error can't find the container with id 9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29 Dec 10 15:08:52 crc kubenswrapper[4847]: I1210 15:08:52.659243 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1240adcf-343f-4c3e-8323-d4ca80e2b99d","Type":"ContainerStarted","Data":"9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29"} Dec 10 15:08:53 crc kubenswrapper[4847]: I1210 15:08:53.684908 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerStarted","Data":"16bb239523b7d803ffb7213a1873f835331604d873ae842644af49510b125792"} Dec 10 15:08:55 crc kubenswrapper[4847]: I1210 15:08:55.712125 4847 generic.go:334] "Generic (PLEG): container finished" podID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerID="16bb239523b7d803ffb7213a1873f835331604d873ae842644af49510b125792" exitCode=0 Dec 10 15:08:55 crc kubenswrapper[4847]: I1210 15:08:55.712392 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerDied","Data":"16bb239523b7d803ffb7213a1873f835331604d873ae842644af49510b125792"} Dec 10 15:09:01 crc kubenswrapper[4847]: I1210 15:09:01.010972 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:09:01 crc kubenswrapper[4847]: I1210 15:09:01.011880 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:09:01 crc kubenswrapper[4847]: I1210 15:09:01.793143 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerStarted","Data":"b41019089f8e8a449e215b77d04fe316b9477450769636e8468ed43841ea79b4"} Dec 10 15:09:10 crc kubenswrapper[4847]: I1210 15:09:10.554033 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:10 crc kubenswrapper[4847]: I1210 15:09:10.554609 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:11 crc kubenswrapper[4847]: I1210 15:09:11.610913 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5t9xh" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="registry-server" probeResult="failure" output=< Dec 10 15:09:11 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 15:09:11 crc kubenswrapper[4847]: > Dec 10 15:09:20 crc kubenswrapper[4847]: I1210 15:09:20.609381 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:20 crc kubenswrapper[4847]: I1210 15:09:20.635604 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5t9xh" podStartSLOduration=21.39268468 podStartE2EDuration="30.635585237s" podCreationTimestamp="2025-12-10 15:08:50 +0000 UTC" firstStartedPulling="2025-12-10 15:08:51.644301176 +0000 UTC m=+2681.213518806" lastFinishedPulling="2025-12-10 15:09:00.887201733 +0000 UTC m=+2690.456419363" observedRunningTime="2025-12-10 15:09:01.823223094 +0000 UTC m=+2691.392440734" watchObservedRunningTime="2025-12-10 15:09:20.635585237 +0000 UTC m=+2710.204802867" Dec 10 15:09:20 crc kubenswrapper[4847]: I1210 15:09:20.673883 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:21 crc kubenswrapper[4847]: I1210 15:09:21.413877 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:09:22 crc kubenswrapper[4847]: I1210 15:09:22.032670 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5t9xh" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="registry-server" containerID="cri-o://b41019089f8e8a449e215b77d04fe316b9477450769636e8468ed43841ea79b4" gracePeriod=2 Dec 10 15:09:24 crc kubenswrapper[4847]: I1210 15:09:24.050201 4847 generic.go:334] "Generic (PLEG): container finished" podID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerID="b41019089f8e8a449e215b77d04fe316b9477450769636e8468ed43841ea79b4" exitCode=0 Dec 10 15:09:24 crc kubenswrapper[4847]: I1210 15:09:24.050248 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerDied","Data":"b41019089f8e8a449e215b77d04fe316b9477450769636e8468ed43841ea79b4"} Dec 10 15:09:25 crc kubenswrapper[4847]: E1210 15:09:25.096926 4847 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 10 15:09:25 crc kubenswrapper[4847]: E1210 15:09:25.097853 4847 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s9thp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1240adcf-343f-4c3e-8323-d4ca80e2b99d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 15:09:25 crc kubenswrapper[4847]: E1210 15:09:25.099455 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.401647 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.545969 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content\") pod \"ac710fa4-2fa9-4913-92b8-85927abf75f5\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.546420 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities\") pod \"ac710fa4-2fa9-4913-92b8-85927abf75f5\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.546517 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khfsx\" (UniqueName: \"kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx\") pod \"ac710fa4-2fa9-4913-92b8-85927abf75f5\" (UID: \"ac710fa4-2fa9-4913-92b8-85927abf75f5\") " Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.547115 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities" (OuterVolumeSpecName: "utilities") pod "ac710fa4-2fa9-4913-92b8-85927abf75f5" (UID: "ac710fa4-2fa9-4913-92b8-85927abf75f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.561459 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx" (OuterVolumeSpecName: "kube-api-access-khfsx") pod "ac710fa4-2fa9-4913-92b8-85927abf75f5" (UID: "ac710fa4-2fa9-4913-92b8-85927abf75f5"). InnerVolumeSpecName "kube-api-access-khfsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.649265 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.649300 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khfsx\" (UniqueName: \"kubernetes.io/projected/ac710fa4-2fa9-4913-92b8-85927abf75f5-kube-api-access-khfsx\") on node \"crc\" DevicePath \"\"" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.652329 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac710fa4-2fa9-4913-92b8-85927abf75f5" (UID: "ac710fa4-2fa9-4913-92b8-85927abf75f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:09:25 crc kubenswrapper[4847]: I1210 15:09:25.751598 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac710fa4-2fa9-4913-92b8-85927abf75f5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.072885 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t9xh" event={"ID":"ac710fa4-2fa9-4913-92b8-85927abf75f5","Type":"ContainerDied","Data":"af30346f6cd7f8b9ae192b32fe64e745cc853bbe36b9da6be63a86fcec3ea636"} Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.072969 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t9xh" Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.072980 4847 scope.go:117] "RemoveContainer" containerID="b41019089f8e8a449e215b77d04fe316b9477450769636e8468ed43841ea79b4" Dec 10 15:09:26 crc kubenswrapper[4847]: E1210 15:09:26.074348 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.115659 4847 scope.go:117] "RemoveContainer" containerID="16bb239523b7d803ffb7213a1873f835331604d873ae842644af49510b125792" Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.130766 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.139429 4847 scope.go:117] "RemoveContainer" containerID="c01eb94a0f840cd8e465de6c351e25cdde5d343d5b6666253681a65afade8a5a" Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.141386 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5t9xh"] Dec 10 15:09:26 crc kubenswrapper[4847]: I1210 15:09:26.793792 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" path="/var/lib/kubelet/pods/ac710fa4-2fa9-4913-92b8-85927abf75f5/volumes" Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.011088 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.011482 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.011525 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.012457 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.012601 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a" gracePeriod=600 Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.145928 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a" exitCode=0 Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.145967 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a"} Dec 10 15:09:31 crc kubenswrapper[4847]: I1210 15:09:31.146017 4847 scope.go:117] "RemoveContainer" containerID="0f81448703741e1d4caf4d16fa3907663a43ae1df8bd0b2c3958f9095044030d" Dec 10 15:09:32 crc kubenswrapper[4847]: I1210 15:09:32.158623 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3"} Dec 10 15:09:41 crc kubenswrapper[4847]: I1210 15:09:41.288189 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 15:09:42 crc kubenswrapper[4847]: I1210 15:09:42.282097 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1240adcf-343f-4c3e-8323-d4ca80e2b99d","Type":"ContainerStarted","Data":"cf4c9139bcf93c7b2579d06290696de5db5df8c1c9623305113ab6528e6689d3"} Dec 10 15:09:42 crc kubenswrapper[4847]: I1210 15:09:42.318889 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.181779042 podStartE2EDuration="52.318862339s" podCreationTimestamp="2025-12-10 15:08:50 +0000 UTC" firstStartedPulling="2025-12-10 15:08:52.148580179 +0000 UTC m=+2681.717797809" lastFinishedPulling="2025-12-10 15:09:41.285663466 +0000 UTC m=+2730.854881106" observedRunningTime="2025-12-10 15:09:42.30436341 +0000 UTC m=+2731.873581080" watchObservedRunningTime="2025-12-10 15:09:42.318862339 +0000 UTC m=+2731.888080009" Dec 10 15:11:31 crc kubenswrapper[4847]: I1210 15:11:31.011151 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:11:31 crc kubenswrapper[4847]: I1210 15:11:31.011846 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:12:01 crc kubenswrapper[4847]: I1210 15:12:01.011683 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:12:01 crc kubenswrapper[4847]: I1210 15:12:01.012365 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.011561 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.012377 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.012444 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.013549 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.013682 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" gracePeriod=600 Dec 10 15:12:31 crc kubenswrapper[4847]: E1210 15:12:31.179309 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.579684 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" exitCode=0 Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.579839 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3"} Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.580304 4847 scope.go:117] "RemoveContainer" containerID="1e7d4473391c32189a95437d015ac3c8a357d44bcf135afee1979d1b74b4076a" Dec 10 15:12:31 crc kubenswrapper[4847]: I1210 15:12:31.581026 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:12:31 crc kubenswrapper[4847]: E1210 15:12:31.581843 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.182276 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:45 crc kubenswrapper[4847]: E1210 15:12:45.183339 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="extract-utilities" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.183360 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="extract-utilities" Dec 10 15:12:45 crc kubenswrapper[4847]: E1210 15:12:45.183394 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="extract-content" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.183403 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="extract-content" Dec 10 15:12:45 crc kubenswrapper[4847]: E1210 15:12:45.183414 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="registry-server" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.183422 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="registry-server" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.183664 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac710fa4-2fa9-4913-92b8-85927abf75f5" containerName="registry-server" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.185546 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.214031 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.359799 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.360518 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gbjx\" (UniqueName: \"kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.360608 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.390642 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.394456 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.407598 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.462400 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.462507 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gbjx\" (UniqueName: \"kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.462563 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.463065 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.463122 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.489685 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gbjx\" (UniqueName: \"kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx\") pod \"redhat-marketplace-tkzdp\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.515434 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.563991 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv8pf\" (UniqueName: \"kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.564066 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.564250 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.666801 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.667563 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv8pf\" (UniqueName: \"kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.667643 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.667646 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.668179 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.707902 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv8pf\" (UniqueName: \"kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf\") pod \"community-operators-889w8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:45 crc kubenswrapper[4847]: I1210 15:12:45.723239 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.039988 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.326461 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:12:46 crc kubenswrapper[4847]: W1210 15:12:46.410414 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod959465d2_2b0e_4492_90a6_61c8970dd6e8.slice/crio-289cba7b5b8ec75a356be84edef9802f2c3d07282523378314f93bbcb56c6088 WatchSource:0}: Error finding container 289cba7b5b8ec75a356be84edef9802f2c3d07282523378314f93bbcb56c6088: Status 404 returned error can't find the container with id 289cba7b5b8ec75a356be84edef9802f2c3d07282523378314f93bbcb56c6088 Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.764042 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:12:46 crc kubenswrapper[4847]: E1210 15:12:46.764345 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.791875 4847 generic.go:334] "Generic (PLEG): container finished" podID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerID="41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270" exitCode=0 Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.791983 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerDied","Data":"41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270"} Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.792041 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerStarted","Data":"0534a4826b0e3adb6447951b452b13aa5c03c35d36e7811110c5955cc021a5dc"} Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.794819 4847 generic.go:334] "Generic (PLEG): container finished" podID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerID="1f3125da8a9212d0b04cb12afd5fa2b85fd80fd7cd21e277e48e3f9e191d3e50" exitCode=0 Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.794852 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerDied","Data":"1f3125da8a9212d0b04cb12afd5fa2b85fd80fd7cd21e277e48e3f9e191d3e50"} Dec 10 15:12:46 crc kubenswrapper[4847]: I1210 15:12:46.794873 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerStarted","Data":"289cba7b5b8ec75a356be84edef9802f2c3d07282523378314f93bbcb56c6088"} Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.585599 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.590826 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.610486 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.706859 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.707239 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmpk9\" (UniqueName: \"kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.707333 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.809234 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerStarted","Data":"c6169b59a0f0e70aecfce81ebec14461dd8a4afcd32f537336db9f7163885a1f"} Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.811500 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.811632 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.811696 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmpk9\" (UniqueName: \"kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.812226 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.812275 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.846937 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmpk9\" (UniqueName: \"kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9\") pod \"certified-operators-wlhjl\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:47 crc kubenswrapper[4847]: I1210 15:12:47.961764 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:48 crc kubenswrapper[4847]: I1210 15:12:48.363743 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:12:48 crc kubenswrapper[4847]: I1210 15:12:48.824644 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerStarted","Data":"1787fa81fd581ddaa3a3a679ab2e6e44e150da26186e6684d41845389bbff54d"} Dec 10 15:12:48 crc kubenswrapper[4847]: I1210 15:12:48.826153 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerStarted","Data":"dc398ec942bde9ccf68ed13f368fea779df3dfafa1011c1812431c831923d570"} Dec 10 15:12:48 crc kubenswrapper[4847]: I1210 15:12:48.829704 4847 generic.go:334] "Generic (PLEG): container finished" podID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerID="f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86" exitCode=0 Dec 10 15:12:48 crc kubenswrapper[4847]: I1210 15:12:48.829932 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerDied","Data":"f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86"} Dec 10 15:12:49 crc kubenswrapper[4847]: I1210 15:12:49.850087 4847 generic.go:334] "Generic (PLEG): container finished" podID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerID="c6169b59a0f0e70aecfce81ebec14461dd8a4afcd32f537336db9f7163885a1f" exitCode=0 Dec 10 15:12:49 crc kubenswrapper[4847]: I1210 15:12:49.850183 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerDied","Data":"c6169b59a0f0e70aecfce81ebec14461dd8a4afcd32f537336db9f7163885a1f"} Dec 10 15:12:49 crc kubenswrapper[4847]: I1210 15:12:49.856421 4847 generic.go:334] "Generic (PLEG): container finished" podID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerID="1787fa81fd581ddaa3a3a679ab2e6e44e150da26186e6684d41845389bbff54d" exitCode=0 Dec 10 15:12:49 crc kubenswrapper[4847]: I1210 15:12:49.856484 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerDied","Data":"1787fa81fd581ddaa3a3a679ab2e6e44e150da26186e6684d41845389bbff54d"} Dec 10 15:12:51 crc kubenswrapper[4847]: I1210 15:12:51.043419 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerStarted","Data":"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef"} Dec 10 15:12:52 crc kubenswrapper[4847]: I1210 15:12:52.065783 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerStarted","Data":"f17c4839450be99c2ef5fe6e98d5287baa51846852c12451e04d50d8839b3fd5"} Dec 10 15:12:52 crc kubenswrapper[4847]: I1210 15:12:52.069412 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerStarted","Data":"cbde111c8b431de6766cbd760019eef206f326defb3a3405ae12afdbc007f7ac"} Dec 10 15:12:52 crc kubenswrapper[4847]: I1210 15:12:52.096061 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-889w8" podStartSLOduration=3.327794838 podStartE2EDuration="7.096035468s" podCreationTimestamp="2025-12-10 15:12:45 +0000 UTC" firstStartedPulling="2025-12-10 15:12:46.796323079 +0000 UTC m=+2916.365540719" lastFinishedPulling="2025-12-10 15:12:50.564563719 +0000 UTC m=+2920.133781349" observedRunningTime="2025-12-10 15:12:52.0873139 +0000 UTC m=+2921.656531540" watchObservedRunningTime="2025-12-10 15:12:52.096035468 +0000 UTC m=+2921.665253108" Dec 10 15:12:52 crc kubenswrapper[4847]: I1210 15:12:52.098372 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tkzdp" podStartSLOduration=4.168893984 podStartE2EDuration="7.098359861s" podCreationTimestamp="2025-12-10 15:12:45 +0000 UTC" firstStartedPulling="2025-12-10 15:12:46.795495277 +0000 UTC m=+2916.364712917" lastFinishedPulling="2025-12-10 15:12:49.724961124 +0000 UTC m=+2919.294178794" observedRunningTime="2025-12-10 15:12:51.083279746 +0000 UTC m=+2920.652497386" watchObservedRunningTime="2025-12-10 15:12:52.098359861 +0000 UTC m=+2921.667577501" Dec 10 15:12:53 crc kubenswrapper[4847]: I1210 15:12:53.085256 4847 generic.go:334] "Generic (PLEG): container finished" podID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerID="cbde111c8b431de6766cbd760019eef206f326defb3a3405ae12afdbc007f7ac" exitCode=0 Dec 10 15:12:53 crc kubenswrapper[4847]: I1210 15:12:53.085333 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerDied","Data":"cbde111c8b431de6766cbd760019eef206f326defb3a3405ae12afdbc007f7ac"} Dec 10 15:12:54 crc kubenswrapper[4847]: I1210 15:12:54.097250 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerStarted","Data":"cd5c4fbeb9d70926a5b33b0c837684006d9c5b4f0964b32eefa353d5af6ee089"} Dec 10 15:12:54 crc kubenswrapper[4847]: I1210 15:12:54.119091 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wlhjl" podStartSLOduration=3.436572807 podStartE2EDuration="7.119074562s" podCreationTimestamp="2025-12-10 15:12:47 +0000 UTC" firstStartedPulling="2025-12-10 15:12:49.859800235 +0000 UTC m=+2919.429017905" lastFinishedPulling="2025-12-10 15:12:53.54230203 +0000 UTC m=+2923.111519660" observedRunningTime="2025-12-10 15:12:54.114207329 +0000 UTC m=+2923.683424959" watchObservedRunningTime="2025-12-10 15:12:54.119074562 +0000 UTC m=+2923.688292192" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.515741 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.516204 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.568862 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.723457 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.723500 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:55 crc kubenswrapper[4847]: I1210 15:12:55.773304 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:56 crc kubenswrapper[4847]: I1210 15:12:56.187617 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:12:56 crc kubenswrapper[4847]: I1210 15:12:56.190468 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:57 crc kubenswrapper[4847]: I1210 15:12:57.962578 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:57 crc kubenswrapper[4847]: I1210 15:12:57.963274 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:58 crc kubenswrapper[4847]: I1210 15:12:58.043978 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:58 crc kubenswrapper[4847]: I1210 15:12:58.184517 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:12:58 crc kubenswrapper[4847]: I1210 15:12:58.587758 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:58 crc kubenswrapper[4847]: I1210 15:12:58.588017 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tkzdp" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="registry-server" containerID="cri-o://c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef" gracePeriod=2 Dec 10 15:12:58 crc kubenswrapper[4847]: I1210 15:12:58.760639 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:12:58 crc kubenswrapper[4847]: E1210 15:12:58.763538 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.056639 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.143610 4847 generic.go:334] "Generic (PLEG): container finished" podID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerID="c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef" exitCode=0 Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.143673 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerDied","Data":"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef"} Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.143755 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tkzdp" event={"ID":"7afbc38c-1674-4ffb-b48a-b36092a6b3b7","Type":"ContainerDied","Data":"0534a4826b0e3adb6447951b452b13aa5c03c35d36e7811110c5955cc021a5dc"} Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.143773 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tkzdp" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.143783 4847 scope.go:117] "RemoveContainer" containerID="c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.170491 4847 scope.go:117] "RemoveContainer" containerID="f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.171784 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content\") pod \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.172373 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities\") pod \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.173000 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities" (OuterVolumeSpecName: "utilities") pod "7afbc38c-1674-4ffb-b48a-b36092a6b3b7" (UID: "7afbc38c-1674-4ffb-b48a-b36092a6b3b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.172404 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gbjx\" (UniqueName: \"kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx\") pod \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\" (UID: \"7afbc38c-1674-4ffb-b48a-b36092a6b3b7\") " Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.173870 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.179116 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx" (OuterVolumeSpecName: "kube-api-access-2gbjx") pod "7afbc38c-1674-4ffb-b48a-b36092a6b3b7" (UID: "7afbc38c-1674-4ffb-b48a-b36092a6b3b7"). InnerVolumeSpecName "kube-api-access-2gbjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.193062 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7afbc38c-1674-4ffb-b48a-b36092a6b3b7" (UID: "7afbc38c-1674-4ffb-b48a-b36092a6b3b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.202006 4847 scope.go:117] "RemoveContainer" containerID="41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.276640 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gbjx\" (UniqueName: \"kubernetes.io/projected/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-kube-api-access-2gbjx\") on node \"crc\" DevicePath \"\"" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.276708 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afbc38c-1674-4ffb-b48a-b36092a6b3b7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.279039 4847 scope.go:117] "RemoveContainer" containerID="c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef" Dec 10 15:12:59 crc kubenswrapper[4847]: E1210 15:12:59.280004 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef\": container with ID starting with c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef not found: ID does not exist" containerID="c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.280102 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef"} err="failed to get container status \"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef\": rpc error: code = NotFound desc = could not find container \"c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef\": container with ID starting with c65d602e423cc8dd7f2d1ddcebba7fac3a4ac7c932cb588260ae7140b1a433ef not found: ID does not exist" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.280134 4847 scope.go:117] "RemoveContainer" containerID="f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86" Dec 10 15:12:59 crc kubenswrapper[4847]: E1210 15:12:59.280585 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86\": container with ID starting with f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86 not found: ID does not exist" containerID="f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.280615 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86"} err="failed to get container status \"f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86\": rpc error: code = NotFound desc = could not find container \"f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86\": container with ID starting with f6ed61b5828356d53d9d635c97b8b21e2d380edeb9db3f9b2db2dc4707637f86 not found: ID does not exist" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.280641 4847 scope.go:117] "RemoveContainer" containerID="41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270" Dec 10 15:12:59 crc kubenswrapper[4847]: E1210 15:12:59.281028 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270\": container with ID starting with 41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270 not found: ID does not exist" containerID="41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.281061 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270"} err="failed to get container status \"41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270\": rpc error: code = NotFound desc = could not find container \"41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270\": container with ID starting with 41adbe46560758ec7d4d1302111e79a611aec6674dd57fc8042103cdfcfe1270 not found: ID does not exist" Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.498153 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.508527 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tkzdp"] Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.974920 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:12:59 crc kubenswrapper[4847]: I1210 15:12:59.975578 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-889w8" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="registry-server" containerID="cri-o://f17c4839450be99c2ef5fe6e98d5287baa51846852c12451e04d50d8839b3fd5" gracePeriod=2 Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.154857 4847 generic.go:334] "Generic (PLEG): container finished" podID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerID="f17c4839450be99c2ef5fe6e98d5287baa51846852c12451e04d50d8839b3fd5" exitCode=0 Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.154899 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerDied","Data":"f17c4839450be99c2ef5fe6e98d5287baa51846852c12451e04d50d8839b3fd5"} Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.502445 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.602453 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content\") pod \"959465d2-2b0e-4492-90a6-61c8970dd6e8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.602596 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv8pf\" (UniqueName: \"kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf\") pod \"959465d2-2b0e-4492-90a6-61c8970dd6e8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.602700 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities\") pod \"959465d2-2b0e-4492-90a6-61c8970dd6e8\" (UID: \"959465d2-2b0e-4492-90a6-61c8970dd6e8\") " Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.603667 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities" (OuterVolumeSpecName: "utilities") pod "959465d2-2b0e-4492-90a6-61c8970dd6e8" (UID: "959465d2-2b0e-4492-90a6-61c8970dd6e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.608799 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf" (OuterVolumeSpecName: "kube-api-access-zv8pf") pod "959465d2-2b0e-4492-90a6-61c8970dd6e8" (UID: "959465d2-2b0e-4492-90a6-61c8970dd6e8"). InnerVolumeSpecName "kube-api-access-zv8pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.665089 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "959465d2-2b0e-4492-90a6-61c8970dd6e8" (UID: "959465d2-2b0e-4492-90a6-61c8970dd6e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.705449 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv8pf\" (UniqueName: \"kubernetes.io/projected/959465d2-2b0e-4492-90a6-61c8970dd6e8-kube-api-access-zv8pf\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.705501 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.705519 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/959465d2-2b0e-4492-90a6-61c8970dd6e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.773470 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" path="/var/lib/kubelet/pods/7afbc38c-1674-4ffb-b48a-b36092a6b3b7/volumes" Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.976674 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:13:00 crc kubenswrapper[4847]: I1210 15:13:00.977292 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wlhjl" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="registry-server" containerID="cri-o://cd5c4fbeb9d70926a5b33b0c837684006d9c5b4f0964b32eefa353d5af6ee089" gracePeriod=2 Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.172086 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-889w8" event={"ID":"959465d2-2b0e-4492-90a6-61c8970dd6e8","Type":"ContainerDied","Data":"289cba7b5b8ec75a356be84edef9802f2c3d07282523378314f93bbcb56c6088"} Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.172155 4847 scope.go:117] "RemoveContainer" containerID="f17c4839450be99c2ef5fe6e98d5287baa51846852c12451e04d50d8839b3fd5" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.172199 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-889w8" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.177553 4847 generic.go:334] "Generic (PLEG): container finished" podID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerID="cd5c4fbeb9d70926a5b33b0c837684006d9c5b4f0964b32eefa353d5af6ee089" exitCode=0 Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.177607 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerDied","Data":"cd5c4fbeb9d70926a5b33b0c837684006d9c5b4f0964b32eefa353d5af6ee089"} Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.200215 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.214169 4847 scope.go:117] "RemoveContainer" containerID="c6169b59a0f0e70aecfce81ebec14461dd8a4afcd32f537336db9f7163885a1f" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.223034 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-889w8"] Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.240697 4847 scope.go:117] "RemoveContainer" containerID="1f3125da8a9212d0b04cb12afd5fa2b85fd80fd7cd21e277e48e3f9e191d3e50" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.516520 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.623940 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities\") pod \"d6f26e11-9e49-4501-9bac-76356eaf4964\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.624070 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmpk9\" (UniqueName: \"kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9\") pod \"d6f26e11-9e49-4501-9bac-76356eaf4964\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.624122 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content\") pod \"d6f26e11-9e49-4501-9bac-76356eaf4964\" (UID: \"d6f26e11-9e49-4501-9bac-76356eaf4964\") " Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.625800 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities" (OuterVolumeSpecName: "utilities") pod "d6f26e11-9e49-4501-9bac-76356eaf4964" (UID: "d6f26e11-9e49-4501-9bac-76356eaf4964"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.634049 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9" (OuterVolumeSpecName: "kube-api-access-vmpk9") pod "d6f26e11-9e49-4501-9bac-76356eaf4964" (UID: "d6f26e11-9e49-4501-9bac-76356eaf4964"). InnerVolumeSpecName "kube-api-access-vmpk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.676487 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6f26e11-9e49-4501-9bac-76356eaf4964" (UID: "d6f26e11-9e49-4501-9bac-76356eaf4964"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.726541 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmpk9\" (UniqueName: \"kubernetes.io/projected/d6f26e11-9e49-4501-9bac-76356eaf4964-kube-api-access-vmpk9\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.726613 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:01 crc kubenswrapper[4847]: I1210 15:13:01.726625 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6f26e11-9e49-4501-9bac-76356eaf4964-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.190455 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlhjl" event={"ID":"d6f26e11-9e49-4501-9bac-76356eaf4964","Type":"ContainerDied","Data":"dc398ec942bde9ccf68ed13f368fea779df3dfafa1011c1812431c831923d570"} Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.190562 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlhjl" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.190868 4847 scope.go:117] "RemoveContainer" containerID="cd5c4fbeb9d70926a5b33b0c837684006d9c5b4f0964b32eefa353d5af6ee089" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.229940 4847 scope.go:117] "RemoveContainer" containerID="cbde111c8b431de6766cbd760019eef206f326defb3a3405ae12afdbc007f7ac" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.240844 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.249707 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wlhjl"] Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.273532 4847 scope.go:117] "RemoveContainer" containerID="1787fa81fd581ddaa3a3a679ab2e6e44e150da26186e6684d41845389bbff54d" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.778430 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" path="/var/lib/kubelet/pods/959465d2-2b0e-4492-90a6-61c8970dd6e8/volumes" Dec 10 15:13:02 crc kubenswrapper[4847]: I1210 15:13:02.779762 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" path="/var/lib/kubelet/pods/d6f26e11-9e49-4501-9bac-76356eaf4964/volumes" Dec 10 15:13:12 crc kubenswrapper[4847]: I1210 15:13:12.759815 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:13:12 crc kubenswrapper[4847]: E1210 15:13:12.760870 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:13:24 crc kubenswrapper[4847]: I1210 15:13:24.760471 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:13:24 crc kubenswrapper[4847]: E1210 15:13:24.761221 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:13:36 crc kubenswrapper[4847]: I1210 15:13:36.762327 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:13:36 crc kubenswrapper[4847]: E1210 15:13:36.763683 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:13:49 crc kubenswrapper[4847]: I1210 15:13:49.759204 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:13:49 crc kubenswrapper[4847]: E1210 15:13:49.759882 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:14:00 crc kubenswrapper[4847]: I1210 15:14:00.775255 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:14:00 crc kubenswrapper[4847]: E1210 15:14:00.776636 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:14:15 crc kubenswrapper[4847]: I1210 15:14:15.759691 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:14:15 crc kubenswrapper[4847]: E1210 15:14:15.760480 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:14:30 crc kubenswrapper[4847]: I1210 15:14:30.777579 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:14:30 crc kubenswrapper[4847]: E1210 15:14:30.778883 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:14:45 crc kubenswrapper[4847]: I1210 15:14:45.760808 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:14:45 crc kubenswrapper[4847]: E1210 15:14:45.762072 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.154654 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br"] Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155663 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155681 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155708 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155729 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155744 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155752 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155768 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155776 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155789 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155796 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155821 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155829 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="extract-utilities" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155854 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155862 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155881 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155889 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.155900 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.155907 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="extract-content" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.156129 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="7afbc38c-1674-4ffb-b48a-b36092a6b3b7" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.156165 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="959465d2-2b0e-4492-90a6-61c8970dd6e8" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.156184 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f26e11-9e49-4501-9bac-76356eaf4964" containerName="registry-server" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.158757 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.160744 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.160830 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.172107 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br"] Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.243333 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlwl\" (UniqueName: \"kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.243421 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.243450 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.345776 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlwl\" (UniqueName: \"kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.345854 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.345878 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.348050 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.353005 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.365654 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlwl\" (UniqueName: \"kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl\") pod \"collect-profiles-29422995-t49br\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.482241 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.765902 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:15:00 crc kubenswrapper[4847]: E1210 15:15:00.766230 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:15:00 crc kubenswrapper[4847]: I1210 15:15:00.954410 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br"] Dec 10 15:15:01 crc kubenswrapper[4847]: I1210 15:15:01.614269 4847 generic.go:334] "Generic (PLEG): container finished" podID="2458b6d4-c3db-4e5e-8b53-5d93f865652f" containerID="ed883318072138e167184360f92a3ebcb217d4bc4f00204dc020ea4291d2283a" exitCode=0 Dec 10 15:15:01 crc kubenswrapper[4847]: I1210 15:15:01.614425 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" event={"ID":"2458b6d4-c3db-4e5e-8b53-5d93f865652f","Type":"ContainerDied","Data":"ed883318072138e167184360f92a3ebcb217d4bc4f00204dc020ea4291d2283a"} Dec 10 15:15:01 crc kubenswrapper[4847]: I1210 15:15:01.614646 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" event={"ID":"2458b6d4-c3db-4e5e-8b53-5d93f865652f","Type":"ContainerStarted","Data":"b1dbec62175811351a6b4cd1f414e3727fffabfd5a4815d70295cfb1b48dab4e"} Dec 10 15:15:02 crc kubenswrapper[4847]: I1210 15:15:02.938728 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.095792 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume\") pod \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.095898 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume\") pod \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.095932 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thlwl\" (UniqueName: \"kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl\") pod \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\" (UID: \"2458b6d4-c3db-4e5e-8b53-5d93f865652f\") " Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.096602 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume" (OuterVolumeSpecName: "config-volume") pod "2458b6d4-c3db-4e5e-8b53-5d93f865652f" (UID: "2458b6d4-c3db-4e5e-8b53-5d93f865652f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.101674 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2458b6d4-c3db-4e5e-8b53-5d93f865652f" (UID: "2458b6d4-c3db-4e5e-8b53-5d93f865652f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.112443 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl" (OuterVolumeSpecName: "kube-api-access-thlwl") pod "2458b6d4-c3db-4e5e-8b53-5d93f865652f" (UID: "2458b6d4-c3db-4e5e-8b53-5d93f865652f"). InnerVolumeSpecName "kube-api-access-thlwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.198081 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2458b6d4-c3db-4e5e-8b53-5d93f865652f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.198760 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2458b6d4-c3db-4e5e-8b53-5d93f865652f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.198801 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thlwl\" (UniqueName: \"kubernetes.io/projected/2458b6d4-c3db-4e5e-8b53-5d93f865652f-kube-api-access-thlwl\") on node \"crc\" DevicePath \"\"" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.633349 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" event={"ID":"2458b6d4-c3db-4e5e-8b53-5d93f865652f","Type":"ContainerDied","Data":"b1dbec62175811351a6b4cd1f414e3727fffabfd5a4815d70295cfb1b48dab4e"} Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.633387 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1dbec62175811351a6b4cd1f414e3727fffabfd5a4815d70295cfb1b48dab4e" Dec 10 15:15:03 crc kubenswrapper[4847]: I1210 15:15:03.633414 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29422995-t49br" Dec 10 15:15:04 crc kubenswrapper[4847]: I1210 15:15:04.011639 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l"] Dec 10 15:15:04 crc kubenswrapper[4847]: I1210 15:15:04.021170 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422950-lmb8l"] Dec 10 15:15:04 crc kubenswrapper[4847]: I1210 15:15:04.771731 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5545dc-5f70-4f7b-950d-077adbf83511" path="/var/lib/kubelet/pods/6c5545dc-5f70-4f7b-950d-077adbf83511/volumes" Dec 10 15:15:13 crc kubenswrapper[4847]: I1210 15:15:13.759853 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:15:13 crc kubenswrapper[4847]: E1210 15:15:13.760556 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:15:24 crc kubenswrapper[4847]: I1210 15:15:24.708977 4847 scope.go:117] "RemoveContainer" containerID="f5488e724c71f37aa03fa183d4c3a16aa08eedd692bbffe68f5d5c47606cd00d" Dec 10 15:15:24 crc kubenswrapper[4847]: I1210 15:15:24.759358 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:15:24 crc kubenswrapper[4847]: E1210 15:15:24.759867 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:15:38 crc kubenswrapper[4847]: I1210 15:15:38.759078 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:15:38 crc kubenswrapper[4847]: E1210 15:15:38.759733 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:15:53 crc kubenswrapper[4847]: I1210 15:15:53.760310 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:15:53 crc kubenswrapper[4847]: E1210 15:15:53.761463 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:16:07 crc kubenswrapper[4847]: I1210 15:16:07.760376 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:16:07 crc kubenswrapper[4847]: E1210 15:16:07.762419 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:16:20 crc kubenswrapper[4847]: I1210 15:16:20.777523 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:16:20 crc kubenswrapper[4847]: E1210 15:16:20.779154 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:16:32 crc kubenswrapper[4847]: I1210 15:16:32.760194 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:16:32 crc kubenswrapper[4847]: E1210 15:16:32.761206 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:16:46 crc kubenswrapper[4847]: I1210 15:16:46.759609 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:16:46 crc kubenswrapper[4847]: E1210 15:16:46.760486 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:16:58 crc kubenswrapper[4847]: I1210 15:16:58.759395 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:16:58 crc kubenswrapper[4847]: E1210 15:16:58.760295 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:17:13 crc kubenswrapper[4847]: I1210 15:17:13.759856 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:17:13 crc kubenswrapper[4847]: E1210 15:17:13.760622 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:17:27 crc kubenswrapper[4847]: I1210 15:17:27.760056 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:17:27 crc kubenswrapper[4847]: E1210 15:17:27.760868 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:17:42 crc kubenswrapper[4847]: I1210 15:17:42.759209 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:17:43 crc kubenswrapper[4847]: I1210 15:17:43.291428 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023"} Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.688470 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:19:57 crc kubenswrapper[4847]: E1210 15:19:57.689643 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2458b6d4-c3db-4e5e-8b53-5d93f865652f" containerName="collect-profiles" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.689662 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2458b6d4-c3db-4e5e-8b53-5d93f865652f" containerName="collect-profiles" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.690128 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="2458b6d4-c3db-4e5e-8b53-5d93f865652f" containerName="collect-profiles" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.691865 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.698624 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.819248 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.819503 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grbt6\" (UniqueName: \"kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.819636 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.921878 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grbt6\" (UniqueName: \"kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.921989 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.922144 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.922807 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.922861 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:57 crc kubenswrapper[4847]: I1210 15:19:57.941566 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grbt6\" (UniqueName: \"kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6\") pod \"redhat-operators-62xsp\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:58 crc kubenswrapper[4847]: I1210 15:19:58.022640 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:19:58 crc kubenswrapper[4847]: I1210 15:19:58.499292 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:19:58 crc kubenswrapper[4847]: I1210 15:19:58.622461 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerStarted","Data":"75d0a3583157da55b6eda6cb1472d1b705d583523387101b468a0785978e578c"} Dec 10 15:19:59 crc kubenswrapper[4847]: I1210 15:19:59.635279 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a9a5477-4379-422f-8e87-ec445020b59b" containerID="467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d" exitCode=0 Dec 10 15:19:59 crc kubenswrapper[4847]: I1210 15:19:59.635406 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerDied","Data":"467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d"} Dec 10 15:19:59 crc kubenswrapper[4847]: I1210 15:19:59.637924 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 15:20:01 crc kubenswrapper[4847]: I1210 15:20:01.011772 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:20:01 crc kubenswrapper[4847]: I1210 15:20:01.012438 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:20:01 crc kubenswrapper[4847]: I1210 15:20:01.660438 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerStarted","Data":"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39"} Dec 10 15:20:03 crc kubenswrapper[4847]: I1210 15:20:03.679266 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a9a5477-4379-422f-8e87-ec445020b59b" containerID="0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39" exitCode=0 Dec 10 15:20:03 crc kubenswrapper[4847]: I1210 15:20:03.679337 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerDied","Data":"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39"} Dec 10 15:20:04 crc kubenswrapper[4847]: I1210 15:20:04.691509 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerStarted","Data":"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c"} Dec 10 15:20:04 crc kubenswrapper[4847]: I1210 15:20:04.720055 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-62xsp" podStartSLOduration=3.079247904 podStartE2EDuration="7.72003052s" podCreationTimestamp="2025-12-10 15:19:57 +0000 UTC" firstStartedPulling="2025-12-10 15:19:59.637626245 +0000 UTC m=+3349.206843875" lastFinishedPulling="2025-12-10 15:20:04.278408861 +0000 UTC m=+3353.847626491" observedRunningTime="2025-12-10 15:20:04.712362631 +0000 UTC m=+3354.281580261" watchObservedRunningTime="2025-12-10 15:20:04.72003052 +0000 UTC m=+3354.289248150" Dec 10 15:20:08 crc kubenswrapper[4847]: I1210 15:20:08.023604 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:08 crc kubenswrapper[4847]: I1210 15:20:08.024390 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:09 crc kubenswrapper[4847]: I1210 15:20:09.074952 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-62xsp" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="registry-server" probeResult="failure" output=< Dec 10 15:20:09 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 15:20:09 crc kubenswrapper[4847]: > Dec 10 15:20:18 crc kubenswrapper[4847]: I1210 15:20:18.098048 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:18 crc kubenswrapper[4847]: I1210 15:20:18.157137 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:18 crc kubenswrapper[4847]: I1210 15:20:18.330531 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:20:19 crc kubenswrapper[4847]: I1210 15:20:19.832791 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-62xsp" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="registry-server" containerID="cri-o://544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c" gracePeriod=2 Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.403921 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.485781 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities\") pod \"8a9a5477-4379-422f-8e87-ec445020b59b\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.486030 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grbt6\" (UniqueName: \"kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6\") pod \"8a9a5477-4379-422f-8e87-ec445020b59b\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.486146 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content\") pod \"8a9a5477-4379-422f-8e87-ec445020b59b\" (UID: \"8a9a5477-4379-422f-8e87-ec445020b59b\") " Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.486698 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities" (OuterVolumeSpecName: "utilities") pod "8a9a5477-4379-422f-8e87-ec445020b59b" (UID: "8a9a5477-4379-422f-8e87-ec445020b59b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.493026 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6" (OuterVolumeSpecName: "kube-api-access-grbt6") pod "8a9a5477-4379-422f-8e87-ec445020b59b" (UID: "8a9a5477-4379-422f-8e87-ec445020b59b"). InnerVolumeSpecName "kube-api-access-grbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.588845 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.588887 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grbt6\" (UniqueName: \"kubernetes.io/projected/8a9a5477-4379-422f-8e87-ec445020b59b-kube-api-access-grbt6\") on node \"crc\" DevicePath \"\"" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.615935 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a9a5477-4379-422f-8e87-ec445020b59b" (UID: "8a9a5477-4379-422f-8e87-ec445020b59b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.690744 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a9a5477-4379-422f-8e87-ec445020b59b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.847890 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a9a5477-4379-422f-8e87-ec445020b59b" containerID="544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c" exitCode=0 Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.847954 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerDied","Data":"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c"} Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.847984 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62xsp" event={"ID":"8a9a5477-4379-422f-8e87-ec445020b59b","Type":"ContainerDied","Data":"75d0a3583157da55b6eda6cb1472d1b705d583523387101b468a0785978e578c"} Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.848000 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62xsp" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.848006 4847 scope.go:117] "RemoveContainer" containerID="544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.876683 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.879107 4847 scope.go:117] "RemoveContainer" containerID="0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.884737 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-62xsp"] Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.905501 4847 scope.go:117] "RemoveContainer" containerID="467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.957253 4847 scope.go:117] "RemoveContainer" containerID="544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c" Dec 10 15:20:20 crc kubenswrapper[4847]: E1210 15:20:20.957787 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c\": container with ID starting with 544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c not found: ID does not exist" containerID="544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.957829 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c"} err="failed to get container status \"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c\": rpc error: code = NotFound desc = could not find container \"544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c\": container with ID starting with 544f20676cbd7ab874237558b4b2df6d44ff0a61cc24ec4a79c66523093bdb3c not found: ID does not exist" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.957858 4847 scope.go:117] "RemoveContainer" containerID="0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39" Dec 10 15:20:20 crc kubenswrapper[4847]: E1210 15:20:20.958079 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39\": container with ID starting with 0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39 not found: ID does not exist" containerID="0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.958106 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39"} err="failed to get container status \"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39\": rpc error: code = NotFound desc = could not find container \"0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39\": container with ID starting with 0a321f118996540f709aa7b9ba5a49bda1483782b1f0940a3e712984eedcfc39 not found: ID does not exist" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.958125 4847 scope.go:117] "RemoveContainer" containerID="467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d" Dec 10 15:20:20 crc kubenswrapper[4847]: E1210 15:20:20.958451 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d\": container with ID starting with 467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d not found: ID does not exist" containerID="467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d" Dec 10 15:20:20 crc kubenswrapper[4847]: I1210 15:20:20.958480 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d"} err="failed to get container status \"467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d\": rpc error: code = NotFound desc = could not find container \"467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d\": container with ID starting with 467eff5e61660ff2e8c7d51ed2e8c783b4b78fb889eeb82f5bc1a824dbcdad7d not found: ID does not exist" Dec 10 15:20:22 crc kubenswrapper[4847]: I1210 15:20:22.770499 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" path="/var/lib/kubelet/pods/8a9a5477-4379-422f-8e87-ec445020b59b/volumes" Dec 10 15:20:31 crc kubenswrapper[4847]: I1210 15:20:31.010772 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:20:31 crc kubenswrapper[4847]: I1210 15:20:31.011302 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.011196 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.011805 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.011844 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.012551 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.012606 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023" gracePeriod=600 Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.261825 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023" exitCode=0 Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.261883 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023"} Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.262458 4847 scope.go:117] "RemoveContainer" containerID="eed995b245f655723093fa61a4d7418fb1eb0ab46a600f56f7264e1150ceffc3" Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.265522 4847 generic.go:334] "Generic (PLEG): container finished" podID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" containerID="cf4c9139bcf93c7b2579d06290696de5db5df8c1c9623305113ab6528e6689d3" exitCode=0 Dec 10 15:21:01 crc kubenswrapper[4847]: I1210 15:21:01.265555 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1240adcf-343f-4c3e-8323-d4ca80e2b99d","Type":"ContainerDied","Data":"cf4c9139bcf93c7b2579d06290696de5db5df8c1c9623305113ab6528e6689d3"} Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.277883 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c"} Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.686638 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.827638 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828014 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828055 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9thp\" (UniqueName: \"kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828077 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828100 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828159 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828279 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828300 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828321 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret\") pod \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\" (UID: \"1240adcf-343f-4c3e-8323-d4ca80e2b99d\") " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.828815 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.829473 4847 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.829767 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data" (OuterVolumeSpecName: "config-data") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.835777 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.838977 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.840990 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp" (OuterVolumeSpecName: "kube-api-access-s9thp") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "kube-api-access-s9thp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.857998 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.858093 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.869018 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.894233 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1240adcf-343f-4c3e-8323-d4ca80e2b99d" (UID: "1240adcf-343f-4c3e-8323-d4ca80e2b99d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932865 4847 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932903 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932918 4847 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1240adcf-343f-4c3e-8323-d4ca80e2b99d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932928 4847 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932940 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9thp\" (UniqueName: \"kubernetes.io/projected/1240adcf-343f-4c3e-8323-d4ca80e2b99d-kube-api-access-s9thp\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932983 4847 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.932995 4847 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1240adcf-343f-4c3e-8323-d4ca80e2b99d-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.933009 4847 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1240adcf-343f-4c3e-8323-d4ca80e2b99d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:02 crc kubenswrapper[4847]: I1210 15:21:02.959875 4847 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 10 15:21:03 crc kubenswrapper[4847]: I1210 15:21:03.034844 4847 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 10 15:21:03 crc kubenswrapper[4847]: I1210 15:21:03.286374 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 15:21:03 crc kubenswrapper[4847]: I1210 15:21:03.286361 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1240adcf-343f-4c3e-8323-d4ca80e2b99d","Type":"ContainerDied","Data":"9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29"} Dec 10 15:21:03 crc kubenswrapper[4847]: I1210 15:21:03.286429 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce0e44a7a5c57af5cd73a5fb556b26647ba70129d8b9e9a548f9877a5e19c29" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.045401 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 15:21:13 crc kubenswrapper[4847]: E1210 15:21:13.047100 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" containerName="tempest-tests-tempest-tests-runner" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047117 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" containerName="tempest-tests-tempest-tests-runner" Dec 10 15:21:13 crc kubenswrapper[4847]: E1210 15:21:13.047133 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="registry-server" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047141 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="registry-server" Dec 10 15:21:13 crc kubenswrapper[4847]: E1210 15:21:13.047293 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="extract-utilities" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047310 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="extract-utilities" Dec 10 15:21:13 crc kubenswrapper[4847]: E1210 15:21:13.047323 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="extract-content" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047331 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="extract-content" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047595 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="1240adcf-343f-4c3e-8323-d4ca80e2b99d" containerName="tempest-tests-tempest-tests-runner" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.047618 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a9a5477-4379-422f-8e87-ec445020b59b" containerName="registry-server" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.048464 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.051518 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-ncmp9" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.056609 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.231469 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.231584 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkqzn\" (UniqueName: \"kubernetes.io/projected/a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7-kube-api-access-tkqzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.333355 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.333430 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkqzn\" (UniqueName: \"kubernetes.io/projected/a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7-kube-api-access-tkqzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.334204 4847 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.355909 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkqzn\" (UniqueName: \"kubernetes.io/projected/a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7-kube-api-access-tkqzn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.364401 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.376543 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 15:21:13 crc kubenswrapper[4847]: I1210 15:21:13.819893 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 15:21:13 crc kubenswrapper[4847]: W1210 15:21:13.828435 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3d5c6e1_891d_4aa5_a9c5_ccaaaee1dad7.slice/crio-cb3939a25c837a3141b9e3051f4c09418e32fc2f040d321f7a7d30b883b3b00b WatchSource:0}: Error finding container cb3939a25c837a3141b9e3051f4c09418e32fc2f040d321f7a7d30b883b3b00b: Status 404 returned error can't find the container with id cb3939a25c837a3141b9e3051f4c09418e32fc2f040d321f7a7d30b883b3b00b Dec 10 15:21:14 crc kubenswrapper[4847]: I1210 15:21:14.383736 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7","Type":"ContainerStarted","Data":"cb3939a25c837a3141b9e3051f4c09418e32fc2f040d321f7a7d30b883b3b00b"} Dec 10 15:21:16 crc kubenswrapper[4847]: I1210 15:21:16.400658 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7","Type":"ContainerStarted","Data":"857e3f571d47ac8c4538a89c188a351632eff2e890cf5d5883564defbd3f38eb"} Dec 10 15:21:16 crc kubenswrapper[4847]: I1210 15:21:16.421191 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.6749804529999999 podStartE2EDuration="3.421171938s" podCreationTimestamp="2025-12-10 15:21:13 +0000 UTC" firstStartedPulling="2025-12-10 15:21:13.83089431 +0000 UTC m=+3423.400111940" lastFinishedPulling="2025-12-10 15:21:15.577085795 +0000 UTC m=+3425.146303425" observedRunningTime="2025-12-10 15:21:16.413857568 +0000 UTC m=+3425.983075208" watchObservedRunningTime="2025-12-10 15:21:16.421171938 +0000 UTC m=+3425.990389568" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.408176 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gk25n/must-gather-lvq8h"] Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.410513 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.418002 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gk25n"/"default-dockercfg-7bhv7" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.418215 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gk25n"/"openshift-service-ca.crt" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.418437 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gk25n"/"kube-root-ca.crt" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.432523 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gk25n/must-gather-lvq8h"] Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.505979 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.506084 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.607826 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.607995 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.608437 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.648141 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm\") pod \"must-gather-lvq8h\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:37 crc kubenswrapper[4847]: I1210 15:21:37.738771 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:21:38 crc kubenswrapper[4847]: I1210 15:21:38.207077 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gk25n/must-gather-lvq8h"] Dec 10 15:21:38 crc kubenswrapper[4847]: I1210 15:21:38.625759 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/must-gather-lvq8h" event={"ID":"91dd8c53-96fb-4766-b8d7-57cd4d199713","Type":"ContainerStarted","Data":"cbfe3bdb39337a160146a28d19582e9dbde82204df7f49ecf336b1902075875f"} Dec 10 15:21:44 crc kubenswrapper[4847]: I1210 15:21:44.702847 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/must-gather-lvq8h" event={"ID":"91dd8c53-96fb-4766-b8d7-57cd4d199713","Type":"ContainerStarted","Data":"3497ea0f03db7cb0d9f27feba5823de6e578e662bbee3d907ebf199aa18c467c"} Dec 10 15:21:45 crc kubenswrapper[4847]: I1210 15:21:45.713793 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/must-gather-lvq8h" event={"ID":"91dd8c53-96fb-4766-b8d7-57cd4d199713","Type":"ContainerStarted","Data":"662af327dc13a4b8333810aed051d1e59b31fbe26c156817922fb9179a59012e"} Dec 10 15:21:45 crc kubenswrapper[4847]: I1210 15:21:45.735327 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gk25n/must-gather-lvq8h" podStartSLOduration=2.5121426529999997 podStartE2EDuration="8.735304309s" podCreationTimestamp="2025-12-10 15:21:37 +0000 UTC" firstStartedPulling="2025-12-10 15:21:38.214384951 +0000 UTC m=+3447.783602581" lastFinishedPulling="2025-12-10 15:21:44.437546607 +0000 UTC m=+3454.006764237" observedRunningTime="2025-12-10 15:21:45.729245354 +0000 UTC m=+3455.298462984" watchObservedRunningTime="2025-12-10 15:21:45.735304309 +0000 UTC m=+3455.304521949" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.226303 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gk25n/crc-debug-8n68d"] Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.228132 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.329848 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.330283 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5k4x\" (UniqueName: \"kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.431972 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5k4x\" (UniqueName: \"kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.432116 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.432281 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.453068 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5k4x\" (UniqueName: \"kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x\") pod \"crc-debug-8n68d\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.546147 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:21:48 crc kubenswrapper[4847]: I1210 15:21:48.737956 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-8n68d" event={"ID":"823a9034-7b9d-4f34-a321-90982c8beecf","Type":"ContainerStarted","Data":"f32cc3f8d6fe8fdf081fe1d1cb4bd1b6bd4402227a45a0765e47cdaa254fa00a"} Dec 10 15:22:01 crc kubenswrapper[4847]: I1210 15:22:01.899674 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-8n68d" event={"ID":"823a9034-7b9d-4f34-a321-90982c8beecf","Type":"ContainerStarted","Data":"8c1a4ece23b596436592e9c879b9e4922967b996c1d57100bf27d060b4813c38"} Dec 10 15:22:01 crc kubenswrapper[4847]: I1210 15:22:01.918337 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gk25n/crc-debug-8n68d" podStartSLOduration=1.5239822159999998 podStartE2EDuration="13.918316859s" podCreationTimestamp="2025-12-10 15:21:48 +0000 UTC" firstStartedPulling="2025-12-10 15:21:48.577758607 +0000 UTC m=+3458.146976237" lastFinishedPulling="2025-12-10 15:22:00.97209325 +0000 UTC m=+3470.541310880" observedRunningTime="2025-12-10 15:22:01.913016204 +0000 UTC m=+3471.482233834" watchObservedRunningTime="2025-12-10 15:22:01.918316859 +0000 UTC m=+3471.487534489" Dec 10 15:22:48 crc kubenswrapper[4847]: I1210 15:22:48.344194 4847 generic.go:334] "Generic (PLEG): container finished" podID="823a9034-7b9d-4f34-a321-90982c8beecf" containerID="8c1a4ece23b596436592e9c879b9e4922967b996c1d57100bf27d060b4813c38" exitCode=0 Dec 10 15:22:48 crc kubenswrapper[4847]: I1210 15:22:48.344287 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-8n68d" event={"ID":"823a9034-7b9d-4f34-a321-90982c8beecf","Type":"ContainerDied","Data":"8c1a4ece23b596436592e9c879b9e4922967b996c1d57100bf27d060b4813c38"} Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.489984 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.527824 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-8n68d"] Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.536790 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-8n68d"] Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.630070 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host\") pod \"823a9034-7b9d-4f34-a321-90982c8beecf\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.630191 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5k4x\" (UniqueName: \"kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x\") pod \"823a9034-7b9d-4f34-a321-90982c8beecf\" (UID: \"823a9034-7b9d-4f34-a321-90982c8beecf\") " Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.630198 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host" (OuterVolumeSpecName: "host") pod "823a9034-7b9d-4f34-a321-90982c8beecf" (UID: "823a9034-7b9d-4f34-a321-90982c8beecf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.630798 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/823a9034-7b9d-4f34-a321-90982c8beecf-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.639474 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x" (OuterVolumeSpecName: "kube-api-access-k5k4x") pod "823a9034-7b9d-4f34-a321-90982c8beecf" (UID: "823a9034-7b9d-4f34-a321-90982c8beecf"). InnerVolumeSpecName "kube-api-access-k5k4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:22:49 crc kubenswrapper[4847]: I1210 15:22:49.732582 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5k4x\" (UniqueName: \"kubernetes.io/projected/823a9034-7b9d-4f34-a321-90982c8beecf-kube-api-access-k5k4x\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.365578 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f32cc3f8d6fe8fdf081fe1d1cb4bd1b6bd4402227a45a0765e47cdaa254fa00a" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.365650 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-8n68d" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.710332 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gk25n/crc-debug-pxp4n"] Dec 10 15:22:50 crc kubenswrapper[4847]: E1210 15:22:50.710830 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="823a9034-7b9d-4f34-a321-90982c8beecf" containerName="container-00" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.710844 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="823a9034-7b9d-4f34-a321-90982c8beecf" containerName="container-00" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.711049 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="823a9034-7b9d-4f34-a321-90982c8beecf" containerName="container-00" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.711701 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.772741 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="823a9034-7b9d-4f34-a321-90982c8beecf" path="/var/lib/kubelet/pods/823a9034-7b9d-4f34-a321-90982c8beecf/volumes" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.857117 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v49wk\" (UniqueName: \"kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.857222 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.959729 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v49wk\" (UniqueName: \"kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.959823 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.960172 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:50 crc kubenswrapper[4847]: I1210 15:22:50.987188 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v49wk\" (UniqueName: \"kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk\") pod \"crc-debug-pxp4n\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:51 crc kubenswrapper[4847]: I1210 15:22:51.030812 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:51 crc kubenswrapper[4847]: I1210 15:22:51.404895 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" event={"ID":"6a97efbb-3796-45fb-b40a-80447ee61337","Type":"ContainerStarted","Data":"9f3831afaa33ac237da49b9d4e74e9d54ba60c1d2635bc873caae279b5983ee7"} Dec 10 15:22:52 crc kubenswrapper[4847]: I1210 15:22:52.415451 4847 generic.go:334] "Generic (PLEG): container finished" podID="6a97efbb-3796-45fb-b40a-80447ee61337" containerID="b8e8ad41c551048b75d5bafdf7489dec80783bf9e55cc85c7d931b3d19ee4c64" exitCode=0 Dec 10 15:22:52 crc kubenswrapper[4847]: I1210 15:22:52.415698 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" event={"ID":"6a97efbb-3796-45fb-b40a-80447ee61337","Type":"ContainerDied","Data":"b8e8ad41c551048b75d5bafdf7489dec80783bf9e55cc85c7d931b3d19ee4c64"} Dec 10 15:22:52 crc kubenswrapper[4847]: I1210 15:22:52.964474 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-pxp4n"] Dec 10 15:22:52 crc kubenswrapper[4847]: I1210 15:22:52.973348 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-pxp4n"] Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.517644 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.624684 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host\") pod \"6a97efbb-3796-45fb-b40a-80447ee61337\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.624774 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v49wk\" (UniqueName: \"kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk\") pod \"6a97efbb-3796-45fb-b40a-80447ee61337\" (UID: \"6a97efbb-3796-45fb-b40a-80447ee61337\") " Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.624889 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host" (OuterVolumeSpecName: "host") pod "6a97efbb-3796-45fb-b40a-80447ee61337" (UID: "6a97efbb-3796-45fb-b40a-80447ee61337"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.625494 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6a97efbb-3796-45fb-b40a-80447ee61337-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.630209 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk" (OuterVolumeSpecName: "kube-api-access-v49wk") pod "6a97efbb-3796-45fb-b40a-80447ee61337" (UID: "6a97efbb-3796-45fb-b40a-80447ee61337"). InnerVolumeSpecName "kube-api-access-v49wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:22:53 crc kubenswrapper[4847]: I1210 15:22:53.727818 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v49wk\" (UniqueName: \"kubernetes.io/projected/6a97efbb-3796-45fb-b40a-80447ee61337-kube-api-access-v49wk\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.135292 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gk25n/crc-debug-v7nxb"] Dec 10 15:22:54 crc kubenswrapper[4847]: E1210 15:22:54.136171 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a97efbb-3796-45fb-b40a-80447ee61337" containerName="container-00" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.136201 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a97efbb-3796-45fb-b40a-80447ee61337" containerName="container-00" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.136411 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a97efbb-3796-45fb-b40a-80447ee61337" containerName="container-00" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.137213 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.339304 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.339371 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8brf\" (UniqueName: \"kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.434183 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f3831afaa33ac237da49b9d4e74e9d54ba60c1d2635bc873caae279b5983ee7" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.434239 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-pxp4n" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.442507 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.442565 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8brf\" (UniqueName: \"kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.442660 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.462033 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8brf\" (UniqueName: \"kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf\") pod \"crc-debug-v7nxb\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.462567 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:54 crc kubenswrapper[4847]: W1210 15:22:54.498789 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05492a3a_e3b2_457d_b021_8bbb1e8c635d.slice/crio-3042308ba5871920d10b0982d2e31d0ad340c346808c2bdaf3defe0eef4a72d7 WatchSource:0}: Error finding container 3042308ba5871920d10b0982d2e31d0ad340c346808c2bdaf3defe0eef4a72d7: Status 404 returned error can't find the container with id 3042308ba5871920d10b0982d2e31d0ad340c346808c2bdaf3defe0eef4a72d7 Dec 10 15:22:54 crc kubenswrapper[4847]: I1210 15:22:54.773039 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a97efbb-3796-45fb-b40a-80447ee61337" path="/var/lib/kubelet/pods/6a97efbb-3796-45fb-b40a-80447ee61337/volumes" Dec 10 15:22:55 crc kubenswrapper[4847]: I1210 15:22:55.444635 4847 generic.go:334] "Generic (PLEG): container finished" podID="05492a3a-e3b2-457d-b021-8bbb1e8c635d" containerID="ec7b3c4b122af3c0ea856b5ae2af14d4c9f1648d4ba6f222a24a1ee886875f04" exitCode=0 Dec 10 15:22:55 crc kubenswrapper[4847]: I1210 15:22:55.444688 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" event={"ID":"05492a3a-e3b2-457d-b021-8bbb1e8c635d","Type":"ContainerDied","Data":"ec7b3c4b122af3c0ea856b5ae2af14d4c9f1648d4ba6f222a24a1ee886875f04"} Dec 10 15:22:55 crc kubenswrapper[4847]: I1210 15:22:55.444745 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" event={"ID":"05492a3a-e3b2-457d-b021-8bbb1e8c635d","Type":"ContainerStarted","Data":"3042308ba5871920d10b0982d2e31d0ad340c346808c2bdaf3defe0eef4a72d7"} Dec 10 15:22:55 crc kubenswrapper[4847]: I1210 15:22:55.485561 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-v7nxb"] Dec 10 15:22:55 crc kubenswrapper[4847]: I1210 15:22:55.495251 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gk25n/crc-debug-v7nxb"] Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.562526 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.584782 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host\") pod \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.585072 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8brf\" (UniqueName: \"kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf\") pod \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\" (UID: \"05492a3a-e3b2-457d-b021-8bbb1e8c635d\") " Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.585889 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host" (OuterVolumeSpecName: "host") pod "05492a3a-e3b2-457d-b021-8bbb1e8c635d" (UID: "05492a3a-e3b2-457d-b021-8bbb1e8c635d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.590746 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf" (OuterVolumeSpecName: "kube-api-access-l8brf") pod "05492a3a-e3b2-457d-b021-8bbb1e8c635d" (UID: "05492a3a-e3b2-457d-b021-8bbb1e8c635d"). InnerVolumeSpecName "kube-api-access-l8brf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.686848 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8brf\" (UniqueName: \"kubernetes.io/projected/05492a3a-e3b2-457d-b021-8bbb1e8c635d-kube-api-access-l8brf\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.687189 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05492a3a-e3b2-457d-b021-8bbb1e8c635d-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:22:56 crc kubenswrapper[4847]: I1210 15:22:56.771511 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05492a3a-e3b2-457d-b021-8bbb1e8c635d" path="/var/lib/kubelet/pods/05492a3a-e3b2-457d-b021-8bbb1e8c635d/volumes" Dec 10 15:22:57 crc kubenswrapper[4847]: I1210 15:22:57.462485 4847 scope.go:117] "RemoveContainer" containerID="ec7b3c4b122af3c0ea856b5ae2af14d4c9f1648d4ba6f222a24a1ee886875f04" Dec 10 15:22:57 crc kubenswrapper[4847]: I1210 15:22:57.462537 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/crc-debug-v7nxb" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.861546 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:00 crc kubenswrapper[4847]: E1210 15:23:00.862603 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05492a3a-e3b2-457d-b021-8bbb1e8c635d" containerName="container-00" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.862620 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="05492a3a-e3b2-457d-b021-8bbb1e8c635d" containerName="container-00" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.862890 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="05492a3a-e3b2-457d-b021-8bbb1e8c635d" containerName="container-00" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.864514 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.874637 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.989109 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dtx\" (UniqueName: \"kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.989201 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:00 crc kubenswrapper[4847]: I1210 15:23:00.989254 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.011685 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.011776 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.091129 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.091329 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74dtx\" (UniqueName: \"kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.091397 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.091607 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.091813 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.112972 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dtx\" (UniqueName: \"kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx\") pod \"community-operators-thsn7\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.183347 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:01 crc kubenswrapper[4847]: I1210 15:23:01.761260 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:02 crc kubenswrapper[4847]: I1210 15:23:02.532324 4847 generic.go:334] "Generic (PLEG): container finished" podID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerID="3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03" exitCode=0 Dec 10 15:23:02 crc kubenswrapper[4847]: I1210 15:23:02.532434 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerDied","Data":"3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03"} Dec 10 15:23:02 crc kubenswrapper[4847]: I1210 15:23:02.532690 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerStarted","Data":"0b2aa399c71df969448f595651dba122f29b22ab8c6f35c19941f36907780dc1"} Dec 10 15:23:03 crc kubenswrapper[4847]: I1210 15:23:03.542513 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerStarted","Data":"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a"} Dec 10 15:23:04 crc kubenswrapper[4847]: I1210 15:23:04.552466 4847 generic.go:334] "Generic (PLEG): container finished" podID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerID="026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a" exitCode=0 Dec 10 15:23:04 crc kubenswrapper[4847]: I1210 15:23:04.552586 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerDied","Data":"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a"} Dec 10 15:23:05 crc kubenswrapper[4847]: I1210 15:23:05.565222 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerStarted","Data":"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12"} Dec 10 15:23:05 crc kubenswrapper[4847]: I1210 15:23:05.588700 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-thsn7" podStartSLOduration=3.126654496 podStartE2EDuration="5.588681524s" podCreationTimestamp="2025-12-10 15:23:00 +0000 UTC" firstStartedPulling="2025-12-10 15:23:02.534233973 +0000 UTC m=+3532.103451593" lastFinishedPulling="2025-12-10 15:23:04.996260991 +0000 UTC m=+3534.565478621" observedRunningTime="2025-12-10 15:23:05.58339571 +0000 UTC m=+3535.152613350" watchObservedRunningTime="2025-12-10 15:23:05.588681524 +0000 UTC m=+3535.157899154" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.184429 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.185634 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.242877 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.669658 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5488ccdfb4-gmxkx_e4d02170-155d-4380-9e77-e8705b47a968/barbican-api/0.log" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.673406 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.726598 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.776935 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5488ccdfb4-gmxkx_e4d02170-155d-4380-9e77-e8705b47a968/barbican-api-log/0.log" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.847455 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9fcb8ccdb-j6kdw_70fb729f-c877-4d76-a5b0-a84470c6c4bc/barbican-keystone-listener/0.log" Dec 10 15:23:11 crc kubenswrapper[4847]: I1210 15:23:11.905707 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9fcb8ccdb-j6kdw_70fb729f-c877-4d76-a5b0-a84470c6c4bc/barbican-keystone-listener-log/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.052744 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b85c87b55-fkkkc_9e5b990d-5058-4827-846b-e5ba776fa0ba/barbican-worker-log/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.055482 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b85c87b55-fkkkc_9e5b990d-5058-4827-846b-e5ba776fa0ba/barbican-worker/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.242829 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/ceilometer-central-agent/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.257297 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm_c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.375642 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/ceilometer-notification-agent/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.426433 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/sg-core/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.449132 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/proxy-httpd/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.620403 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f0c3090a-89da-46f1-9998-a68188814b05/cinder-api/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.659917 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f0c3090a-89da-46f1-9998-a68188814b05/cinder-api-log/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.744161 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6b8b833f-7fa3-4232-831a-18d4ba96304b/cinder-scheduler/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.910381 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6b8b833f-7fa3-4232-831a-18d4ba96304b/probe/0.log" Dec 10 15:23:12 crc kubenswrapper[4847]: I1210 15:23:12.994151 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf_21486092-46dc-4bf8-9cfc-5fc5781461c2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.085403 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7_da983cff-081c-4811-9ed3-f3cc109489bd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.188874 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/init/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.400426 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/init/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.503263 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/dnsmasq-dns/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.508638 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-krj6h_48c5aa38-4902-4105-99ac-40fb8ab5a288/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.635598 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-thsn7" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="registry-server" containerID="cri-o://9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12" gracePeriod=2 Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.667804 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23c6ec92-8230-4322-ad9c-67c0a3260eec/glance-httpd/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.680843 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23c6ec92-8230-4322-ad9c-67c0a3260eec/glance-log/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.886068 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4454a179-0fc2-4623-b1ba-b7946009d769/glance-httpd/0.log" Dec 10 15:23:13 crc kubenswrapper[4847]: I1210 15:23:13.887831 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4454a179-0fc2-4623-b1ba-b7946009d769/glance-log/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.100731 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4f6b7576-wvhs4_f7e7178d-cd26-4623-a05c-6eee95cb5dd6/horizon/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.125308 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.174919 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74dtx\" (UniqueName: \"kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx\") pod \"d4d2085f-74d3-46c1-9994-56c64f04eed2\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.175064 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content\") pod \"d4d2085f-74d3-46c1-9994-56c64f04eed2\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.175108 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities\") pod \"d4d2085f-74d3-46c1-9994-56c64f04eed2\" (UID: \"d4d2085f-74d3-46c1-9994-56c64f04eed2\") " Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.176694 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities" (OuterVolumeSpecName: "utilities") pod "d4d2085f-74d3-46c1-9994-56c64f04eed2" (UID: "d4d2085f-74d3-46c1-9994-56c64f04eed2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.182821 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx" (OuterVolumeSpecName: "kube-api-access-74dtx") pod "d4d2085f-74d3-46c1-9994-56c64f04eed2" (UID: "d4d2085f-74d3-46c1-9994-56c64f04eed2"). InnerVolumeSpecName "kube-api-access-74dtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.239271 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4d2085f-74d3-46c1-9994-56c64f04eed2" (UID: "d4d2085f-74d3-46c1-9994-56c64f04eed2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.249376 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jlctv_a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.276747 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.276791 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4d2085f-74d3-46c1-9994-56c64f04eed2-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.276802 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74dtx\" (UniqueName: \"kubernetes.io/projected/d4d2085f-74d3-46c1-9994-56c64f04eed2-kube-api-access-74dtx\") on node \"crc\" DevicePath \"\"" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.369445 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4f6b7576-wvhs4_f7e7178d-cd26-4623-a05c-6eee95cb5dd6/horizon-log/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.387397 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hvd6g_d4c9d846-3adb-4f21-8945-f27d14123b1e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.655313 4847 generic.go:334] "Generic (PLEG): container finished" podID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerID="9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12" exitCode=0 Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.655366 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerDied","Data":"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12"} Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.655398 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-thsn7" event={"ID":"d4d2085f-74d3-46c1-9994-56c64f04eed2","Type":"ContainerDied","Data":"0b2aa399c71df969448f595651dba122f29b22ab8c6f35c19941f36907780dc1"} Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.655429 4847 scope.go:117] "RemoveContainer" containerID="9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.655600 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-thsn7" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.680694 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29422981-q652q_b0790548-d5d8-4688-9120-bb942f395fd2/keystone-cron/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.690193 4847 scope.go:117] "RemoveContainer" containerID="026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.702284 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54599d665c-kl29l_8f2f4d1f-1fc9-48cd-8697-bdc788483a1c/keystone-api/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.710538 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.723655 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-thsn7"] Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.726942 4847 scope.go:117] "RemoveContainer" containerID="3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.780412 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" path="/var/lib/kubelet/pods/d4d2085f-74d3-46c1-9994-56c64f04eed2/volumes" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.782898 4847 scope.go:117] "RemoveContainer" containerID="9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12" Dec 10 15:23:14 crc kubenswrapper[4847]: E1210 15:23:14.783450 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12\": container with ID starting with 9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12 not found: ID does not exist" containerID="9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.783498 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12"} err="failed to get container status \"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12\": rpc error: code = NotFound desc = could not find container \"9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12\": container with ID starting with 9527df39fe12a70820cc73ca4129f28a486a751d24f3b83414bdf9d086d81c12 not found: ID does not exist" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.783540 4847 scope.go:117] "RemoveContainer" containerID="026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a" Dec 10 15:23:14 crc kubenswrapper[4847]: E1210 15:23:14.786935 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a\": container with ID starting with 026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a not found: ID does not exist" containerID="026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.787365 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a"} err="failed to get container status \"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a\": rpc error: code = NotFound desc = could not find container \"026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a\": container with ID starting with 026ac8939b1592cc589af4f328f9d6d2a303216d356ae6bf6d1d5d09388a136a not found: ID does not exist" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.787402 4847 scope.go:117] "RemoveContainer" containerID="3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03" Dec 10 15:23:14 crc kubenswrapper[4847]: E1210 15:23:14.794852 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03\": container with ID starting with 3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03 not found: ID does not exist" containerID="3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.794901 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03"} err="failed to get container status \"3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03\": rpc error: code = NotFound desc = could not find container \"3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03\": container with ID starting with 3d3a5c436feedf07d11fce54f17de60c3135ed01af739fdcc6f25bc76e711a03 not found: ID does not exist" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.918985 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6511a21a-53cd-4ecc-a550-cc4a3ef72c8c/kube-state-metrics/0.log" Dec 10 15:23:14 crc kubenswrapper[4847]: I1210 15:23:14.967769 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp_eb5a7f57-0eb5-47f9-ae7c-173f849c7a02/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:15 crc kubenswrapper[4847]: I1210 15:23:15.358920 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c86c8879c-drvf5_7d85a520-fbc1-4a46-b33e-7513646bfeaf/neutron-httpd/0.log" Dec 10 15:23:15 crc kubenswrapper[4847]: I1210 15:23:15.417534 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c86c8879c-drvf5_7d85a520-fbc1-4a46-b33e-7513646bfeaf/neutron-api/0.log" Dec 10 15:23:15 crc kubenswrapper[4847]: I1210 15:23:15.584503 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c_04d0cfc8-2361-4e74-bacc-379a98054d97/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:15 crc kubenswrapper[4847]: I1210 15:23:15.998153 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e9315ad6-b2ce-420a-b8a4-c168264662de/nova-cell0-conductor-conductor/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.027091 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0205d3f1-fa73-4291-a22b-1bdd4aaa23ec/nova-api-log/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.174233 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0205d3f1-fa73-4291-a22b-1bdd4aaa23ec/nova-api-api/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.278055 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1598cbc7-d655-494f-a2b1-ddbd4db56aa5/nova-cell1-conductor-conductor/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.379573 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d1d12a43-2aba-411a-ad3d-628d6f468c95/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.625437 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ct6mn_9b520035-682d-43ae-acba-403455d78c9f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:16 crc kubenswrapper[4847]: I1210 15:23:16.697754 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d7f8698b-dd55-4811-81da-55dac46eeffe/nova-metadata-log/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.092112 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/mysql-bootstrap/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.134540 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8b228613-6a71-4c3b-a95a-e0a496c0b346/nova-scheduler-scheduler/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.306258 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/mysql-bootstrap/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.315732 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/galera/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.515245 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/mysql-bootstrap/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.715318 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/galera/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.734189 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/mysql-bootstrap/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.884621 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2559a413-0d27-43f0-8fa3-2da6e9d66eff/openstackclient/0.log" Dec 10 15:23:17 crc kubenswrapper[4847]: I1210 15:23:17.937391 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2xjh7_b962e42f-e21b-4016-9587-111fb6eb68ac/ovn-controller/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.118641 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d7f8698b-dd55-4811-81da-55dac46eeffe/nova-metadata-metadata/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.158725 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-n4rbc_4f080a9d-d07e-4a5e-b782-245d4c7b31cf/openstack-network-exporter/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.298766 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server-init/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.484262 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovs-vswitchd/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.490278 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.534908 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server-init/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.739108 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vsngf_dfed455d-08ba-417a-975b-09dc84eaec54/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.746195 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a4d0db06-c1bd-4af6-b78e-a1703d914c6d/ovn-northd/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.793923 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a4d0db06-c1bd-4af6-b78e-a1703d914c6d/openstack-network-exporter/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.927196 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_26654ec0-c7fa-44da-9be9-8d5218addac5/openstack-network-exporter/0.log" Dec 10 15:23:18 crc kubenswrapper[4847]: I1210 15:23:18.960036 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_26654ec0-c7fa-44da-9be9-8d5218addac5/ovsdbserver-nb/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.212321 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_beaec882-8047-4348-bcf5-a45eed4dba20/openstack-network-exporter/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.231646 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_beaec882-8047-4348-bcf5-a45eed4dba20/ovsdbserver-sb/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.463768 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d896bd648-rk9qk_7895251f-c0bb-40b7-a4a2-802a3517cd98/placement-api/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.500751 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d896bd648-rk9qk_7895251f-c0bb-40b7-a4a2-802a3517cd98/placement-log/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.538848 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/setup-container/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.776699 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/setup-container/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.781803 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/rabbitmq/0.log" Dec 10 15:23:19 crc kubenswrapper[4847]: I1210 15:23:19.845488 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/setup-container/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.061344 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/rabbitmq/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.072515 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/setup-container/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.091481 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr_4ef271cb-44e6-4f26-95fc-3ca47f6ab290/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.313128 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s87nz_d39485c7-eaab-4829-b978-ca8109f82227/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.328624 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd_1f79a99b-09b0-4951-bf05-42fe68d44e53/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.559699 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xs56g_f9991c0e-7804-406a-bc24-92c5d5d7fa61/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.591545 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-4ckkd_6e914d68-6bc2-420a-b0cb-3f424d71652c/ssh-known-hosts-edpm-deployment/0.log" Dec 10 15:23:20 crc kubenswrapper[4847]: I1210 15:23:20.967533 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65d99c95b7-t6kg4_9cd88676-59ca-4c55-aaf7-caab8965d2bb/proxy-httpd/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.033255 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65d99c95b7-t6kg4_9cd88676-59ca-4c55-aaf7-caab8965d2bb/proxy-server/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.111381 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hkv2r_daaeae82-74d4-4d18-8178-58ca7246b605/swift-ring-rebalance/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.233495 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-reaper/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.287140 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-auditor/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.329616 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-replicator/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.409197 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-server/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.482130 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-auditor/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.560740 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-server/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.625684 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-updater/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.752202 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-expirer/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.852507 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-replicator/0.log" Dec 10 15:23:21 crc kubenswrapper[4847]: I1210 15:23:21.960357 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-server/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.020372 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-updater/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.154873 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/rsync/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.221522 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/swift-recon-cron/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.333570 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-auditor/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.340543 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-replicator/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.474907 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp_f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.565451 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1240adcf-343f-4c3e-8323-d4ca80e2b99d/tempest-tests-tempest-tests-runner/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.676531 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7/test-operator-logs-container/0.log" Dec 10 15:23:22 crc kubenswrapper[4847]: I1210 15:23:22.778486 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr_7a627a8d-ac7d-4d5d-9258-3342ca59930d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:23:30 crc kubenswrapper[4847]: I1210 15:23:30.834977 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fc672732-248d-4f66-88fd-3b98ce15e78d/memcached/0.log" Dec 10 15:23:31 crc kubenswrapper[4847]: I1210 15:23:31.011168 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:23:31 crc kubenswrapper[4847]: I1210 15:23:31.011218 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.403449 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.638038 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.640414 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.668455 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.846244 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/extract/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.847309 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:23:46 crc kubenswrapper[4847]: I1210 15:23:46.849644 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.043743 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqjxj_80cfb202-f181-4c1a-b9d6-a7cd036b93db/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.112294 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqjxj_80cfb202-f181-4c1a-b9d6-a7cd036b93db/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.171067 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-pdp9r_df30724c-6ddb-4e68-8f57-31f8bb262678/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.293286 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-pdp9r_df30724c-6ddb-4e68-8f57-31f8bb262678/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.346416 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-pmnjz_3c7abfb3-9360-42df-86b1-127066817fe8/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.383209 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-pmnjz_3c7abfb3-9360-42df-86b1-127066817fe8/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.554137 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-fl8dl_519d2b7e-57ab-4f4e-a199-68b3fe9a16bf/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.572885 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-fl8dl_519d2b7e-57ab-4f4e-a199-68b3fe9a16bf/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.674066 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-t2wb8_defc8bda-15a6-4432-8079-9606a8fdb2e9/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.771383 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-t2wb8_defc8bda-15a6-4432-8079-9606a8fdb2e9/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.834515 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gdhjq_82fa6e27-3c61-4a46-a51c-1f9e034a870c/kube-rbac-proxy/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.869987 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gdhjq_82fa6e27-3c61-4a46-a51c-1f9e034a870c/manager/0.log" Dec 10 15:23:47 crc kubenswrapper[4847]: I1210 15:23:47.949179 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-twm4t_73f1dc78-cf06-4c8d-a57a-314ce95cc1d5/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.165333 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4j2fk_3a197914-3847-44b5-aa44-527bef265cc5/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.224895 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4j2fk_3a197914-3847-44b5-aa44-527bef265cc5/manager/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.284745 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-twm4t_73f1dc78-cf06-4c8d-a57a-314ce95cc1d5/manager/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.348427 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-55zkq_2fa5c5a1-e3d6-424c-b52a-d01af15458cf/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.455609 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-55zkq_2fa5c5a1-e3d6-424c-b52a-d01af15458cf/manager/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.519305 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-q8fkp_161aaad4-8b14-4a9c-b793-a7937e078aef/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.571400 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-q8fkp_161aaad4-8b14-4a9c-b793-a7937e078aef/manager/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.674460 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-sptmg_9f72b089-2518-4144-a558-bfb5c490d594/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.713643 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-sptmg_9f72b089-2518-4144-a558-bfb5c490d594/manager/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.852414 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gjtlq_3654ef6c-9eef-40ec-acd3-c101a817f345/kube-rbac-proxy/0.log" Dec 10 15:23:48 crc kubenswrapper[4847]: I1210 15:23:48.900109 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gjtlq_3654ef6c-9eef-40ec-acd3-c101a817f345/manager/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.026727 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-kds4v_79090a78-3da6-4b17-a875-fc8c2d534076/kube-rbac-proxy/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.101038 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-kds4v_79090a78-3da6-4b17-a875-fc8c2d534076/manager/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.128002 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c96c4_db714bfd-5854-4a09-a6d0-acf7e3c539a4/kube-rbac-proxy/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.236070 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c96c4_db714bfd-5854-4a09-a6d0-acf7e3c539a4/manager/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.326684 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f2swnv_d78cab92-e99d-41fb-a35d-ff63d3ceba23/kube-rbac-proxy/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.340977 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f2swnv_d78cab92-e99d-41fb-a35d-ff63d3ceba23/manager/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.843569 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6646c79bbd-rqsnh_caaff4fe-5b58-4a77-b3fe-217742051274/operator/0.log" Dec 10 15:23:49 crc kubenswrapper[4847]: I1210 15:23:49.845496 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ptjmg_692f8962-1682-4a0d-b693-24e90e81bb17/registry-server/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.094605 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sbtv2_64a44edf-788e-4286-83af-6072b8d8b437/kube-rbac-proxy/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.532777 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sbtv2_64a44edf-788e-4286-83af-6072b8d8b437/manager/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.657031 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-t224h_87dac2ce-ec87-4d12-9ac8-0a79bedbfa62/kube-rbac-proxy/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.740293 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-d5mdv_4706d44c-19de-49a6-b604-dbe911592e48/operator/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.867985 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-t224h_87dac2ce-ec87-4d12-9ac8-0a79bedbfa62/manager/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.957340 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-4dfs5_9b0a2f2c-971b-40cc-8544-9842f7878d62/kube-rbac-proxy/0.log" Dec 10 15:23:50 crc kubenswrapper[4847]: I1210 15:23:50.980073 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-4dfs5_9b0a2f2c-971b-40cc-8544-9842f7878d62/manager/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.078902 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79b6c4fddb-8mzvt_b1d011f5-5052-431d-9933-314f65930ce4/manager/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.142974 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-z6f9c_971c05b8-38ac-411a-bdc8-b40ed70daab5/kube-rbac-proxy/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.254523 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-z6f9c_971c05b8-38ac-411a-bdc8-b40ed70daab5/manager/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.296412 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8k92s_c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90/kube-rbac-proxy/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.380073 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8k92s_c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90/manager/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.449318 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-4dhdf_0c5922f3-c544-4993-b978-4a035ec08e1b/kube-rbac-proxy/0.log" Dec 10 15:23:51 crc kubenswrapper[4847]: I1210 15:23:51.503510 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-4dhdf_0c5922f3-c544-4993-b978-4a035ec08e1b/manager/0.log" Dec 10 15:24:01 crc kubenswrapper[4847]: I1210 15:24:01.011032 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:24:01 crc kubenswrapper[4847]: I1210 15:24:01.011520 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:24:01 crc kubenswrapper[4847]: I1210 15:24:01.011565 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:24:01 crc kubenswrapper[4847]: I1210 15:24:01.012236 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:24:01 crc kubenswrapper[4847]: I1210 15:24:01.012277 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" gracePeriod=600 Dec 10 15:24:01 crc kubenswrapper[4847]: E1210 15:24:01.146329 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:24:02 crc kubenswrapper[4847]: I1210 15:24:02.099918 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" exitCode=0 Dec 10 15:24:02 crc kubenswrapper[4847]: I1210 15:24:02.100321 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c"} Dec 10 15:24:02 crc kubenswrapper[4847]: I1210 15:24:02.100385 4847 scope.go:117] "RemoveContainer" containerID="e4f5fb149078971274a4e7adf8e63fc2f16133df7210d25c711131cba31d5023" Dec 10 15:24:02 crc kubenswrapper[4847]: I1210 15:24:02.101287 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:24:02 crc kubenswrapper[4847]: E1210 15:24:02.101918 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:24:08 crc kubenswrapper[4847]: I1210 15:24:08.905295 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-wx4gj_89217e6c-9f20-45f0-8a26-7c808efc3359/control-plane-machine-set-operator/0.log" Dec 10 15:24:09 crc kubenswrapper[4847]: I1210 15:24:09.106015 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6lxgl_57fb36d2-42fe-4993-a454-1db53581dc35/machine-api-operator/0.log" Dec 10 15:24:09 crc kubenswrapper[4847]: I1210 15:24:09.110615 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6lxgl_57fb36d2-42fe-4993-a454-1db53581dc35/kube-rbac-proxy/0.log" Dec 10 15:24:16 crc kubenswrapper[4847]: I1210 15:24:16.760431 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:24:16 crc kubenswrapper[4847]: E1210 15:24:16.761145 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:24:20 crc kubenswrapper[4847]: I1210 15:24:20.804389 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6nfcw_f0cacf0f-d6c7-45c1-9241-7a3eb577594c/cert-manager-controller/0.log" Dec 10 15:24:20 crc kubenswrapper[4847]: I1210 15:24:20.983059 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g8v52_ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d/cert-manager-cainjector/0.log" Dec 10 15:24:21 crc kubenswrapper[4847]: I1210 15:24:21.046357 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7h9v6_2266f4a5-b282-4686-9a2c-e51d538a11cd/cert-manager-webhook/0.log" Dec 10 15:24:28 crc kubenswrapper[4847]: I1210 15:24:28.760508 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:24:28 crc kubenswrapper[4847]: E1210 15:24:28.761470 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:24:32 crc kubenswrapper[4847]: I1210 15:24:32.577044 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-tjd2c_d5575bcb-0d29-4538-8213-99a007548880/nmstate-console-plugin/0.log" Dec 10 15:24:32 crc kubenswrapper[4847]: I1210 15:24:32.705622 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wxpbs_b51a4da1-a921-414f-9985-044264a489ad/nmstate-handler/0.log" Dec 10 15:24:32 crc kubenswrapper[4847]: I1210 15:24:32.729616 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-5rj2v_e7bb30c3-5f7b-465a-8852-388f8204277b/kube-rbac-proxy/0.log" Dec 10 15:24:32 crc kubenswrapper[4847]: I1210 15:24:32.808399 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-5rj2v_e7bb30c3-5f7b-465a-8852-388f8204277b/nmstate-metrics/0.log" Dec 10 15:24:32 crc kubenswrapper[4847]: I1210 15:24:32.965832 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-x4fhc_64363456-a381-49ba-98f2-b16b9928b65b/nmstate-operator/0.log" Dec 10 15:24:33 crc kubenswrapper[4847]: I1210 15:24:33.028390 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4kbcp_ef0ffd00-af07-4843-9b34-2cf613252d7c/nmstate-webhook/0.log" Dec 10 15:24:39 crc kubenswrapper[4847]: I1210 15:24:39.760139 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:24:39 crc kubenswrapper[4847]: E1210 15:24:39.760900 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:24:46 crc kubenswrapper[4847]: I1210 15:24:46.620648 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6mjl9_d00cc0ff-5b61-4786-abd8-44da90bdb484/kube-rbac-proxy/0.log" Dec 10 15:24:46 crc kubenswrapper[4847]: I1210 15:24:46.682411 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6mjl9_d00cc0ff-5b61-4786-abd8-44da90bdb484/controller/0.log" Dec 10 15:24:46 crc kubenswrapper[4847]: I1210 15:24:46.791033 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:24:46 crc kubenswrapper[4847]: I1210 15:24:46.969478 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:24:46 crc kubenswrapper[4847]: I1210 15:24:46.973943 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.001026 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.018567 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.197839 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.214752 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.246805 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.247231 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.399354 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.409209 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.423840 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.437352 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/controller/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.596871 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/frr-metrics/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.629022 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/kube-rbac-proxy/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.656524 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/kube-rbac-proxy-frr/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.815832 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/reloader/0.log" Dec 10 15:24:47 crc kubenswrapper[4847]: I1210 15:24:47.945066 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-ng6j2_ec15bd80-1069-4aed-9f51-43140a7f6c5e/frr-k8s-webhook-server/0.log" Dec 10 15:24:48 crc kubenswrapper[4847]: I1210 15:24:48.112074 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b5f976d54-lrrhd_12bcc429-a2dd-4b92-a99b-86ea695e6147/manager/0.log" Dec 10 15:24:48 crc kubenswrapper[4847]: I1210 15:24:48.253591 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7bc66b7c4c-gxscv_528aeb2a-2b6f-4e7b-8a22-b40ecca39c45/webhook-server/0.log" Dec 10 15:24:48 crc kubenswrapper[4847]: I1210 15:24:48.392827 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7bf6v_74b7d233-5faf-4997-9986-f6fb167df94b/kube-rbac-proxy/0.log" Dec 10 15:24:48 crc kubenswrapper[4847]: I1210 15:24:48.951181 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7bf6v_74b7d233-5faf-4997-9986-f6fb167df94b/speaker/0.log" Dec 10 15:24:49 crc kubenswrapper[4847]: I1210 15:24:49.001821 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/frr/0.log" Dec 10 15:24:50 crc kubenswrapper[4847]: I1210 15:24:50.768331 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:24:50 crc kubenswrapper[4847]: E1210 15:24:50.768997 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.033729 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.247878 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.302081 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.319679 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.493557 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.497237 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/extract/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.505574 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.700534 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.856516 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.858211 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:25:01 crc kubenswrapper[4847]: I1210 15:25:01.901135 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.099415 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.099836 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.153505 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/extract/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.265839 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.420072 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.474876 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.481427 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.630159 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.666275 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.829224 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:25:02 crc kubenswrapper[4847]: I1210 15:25:02.877958 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/registry-server/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.014174 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.038415 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.062959 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.189121 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.196070 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.433402 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dz9r4_f55386e3-acde-49bd-941e-a92591ec95de/marketplace-operator/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.587631 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.683527 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/registry-server/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.697660 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.779857 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.802514 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.930251 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:25:03 crc kubenswrapper[4847]: I1210 15:25:03.949083 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.118871 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/registry-server/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.197676 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.309267 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.353633 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.369085 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.554127 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.565272 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:25:04 crc kubenswrapper[4847]: I1210 15:25:04.928635 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/registry-server/0.log" Dec 10 15:25:05 crc kubenswrapper[4847]: I1210 15:25:05.760230 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:25:05 crc kubenswrapper[4847]: E1210 15:25:05.760767 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:25:20 crc kubenswrapper[4847]: I1210 15:25:20.764999 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:25:20 crc kubenswrapper[4847]: E1210 15:25:20.765670 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:25:32 crc kubenswrapper[4847]: I1210 15:25:32.759542 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:25:32 crc kubenswrapper[4847]: E1210 15:25:32.760480 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:25:45 crc kubenswrapper[4847]: I1210 15:25:45.760415 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:25:45 crc kubenswrapper[4847]: E1210 15:25:45.761246 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:00 crc kubenswrapper[4847]: I1210 15:26:00.765232 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:26:00 crc kubenswrapper[4847]: E1210 15:26:00.766063 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:15 crc kubenswrapper[4847]: I1210 15:26:15.760057 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:26:15 crc kubenswrapper[4847]: E1210 15:26:15.760735 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:15 crc kubenswrapper[4847]: I1210 15:26:15.804840 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3cd44098-3d49-408d-8478-fceaa9ac7e57" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 10 15:26:15 crc kubenswrapper[4847]: I1210 15:26:15.805294 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3cd44098-3d49-408d-8478-fceaa9ac7e57" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 10 15:26:30 crc kubenswrapper[4847]: I1210 15:26:30.761093 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:26:30 crc kubenswrapper[4847]: E1210 15:26:30.761863 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:43 crc kubenswrapper[4847]: I1210 15:26:43.761499 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:26:43 crc kubenswrapper[4847]: E1210 15:26:43.762516 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:51 crc kubenswrapper[4847]: I1210 15:26:51.767371 4847 generic.go:334] "Generic (PLEG): container finished" podID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerID="3497ea0f03db7cb0d9f27feba5823de6e578e662bbee3d907ebf199aa18c467c" exitCode=0 Dec 10 15:26:51 crc kubenswrapper[4847]: I1210 15:26:51.767460 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gk25n/must-gather-lvq8h" event={"ID":"91dd8c53-96fb-4766-b8d7-57cd4d199713","Type":"ContainerDied","Data":"3497ea0f03db7cb0d9f27feba5823de6e578e662bbee3d907ebf199aa18c467c"} Dec 10 15:26:51 crc kubenswrapper[4847]: I1210 15:26:51.769599 4847 scope.go:117] "RemoveContainer" containerID="3497ea0f03db7cb0d9f27feba5823de6e578e662bbee3d907ebf199aa18c467c" Dec 10 15:26:52 crc kubenswrapper[4847]: I1210 15:26:52.160582 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gk25n_must-gather-lvq8h_91dd8c53-96fb-4766-b8d7-57cd4d199713/gather/0.log" Dec 10 15:26:57 crc kubenswrapper[4847]: I1210 15:26:57.760547 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:26:57 crc kubenswrapper[4847]: E1210 15:26:57.761317 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.492130 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gk25n/must-gather-lvq8h"] Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.492733 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gk25n/must-gather-lvq8h" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="copy" containerID="cri-o://662af327dc13a4b8333810aed051d1e59b31fbe26c156817922fb9179a59012e" gracePeriod=2 Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.501108 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gk25n/must-gather-lvq8h"] Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.863178 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gk25n_must-gather-lvq8h_91dd8c53-96fb-4766-b8d7-57cd4d199713/copy/0.log" Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.864040 4847 generic.go:334] "Generic (PLEG): container finished" podID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerID="662af327dc13a4b8333810aed051d1e59b31fbe26c156817922fb9179a59012e" exitCode=143 Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.864085 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbfe3bdb39337a160146a28d19582e9dbde82204df7f49ecf336b1902075875f" Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.930438 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gk25n_must-gather-lvq8h_91dd8c53-96fb-4766-b8d7-57cd4d199713/copy/0.log" Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.930780 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.967549 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output\") pod \"91dd8c53-96fb-4766-b8d7-57cd4d199713\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.967610 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm\") pod \"91dd8c53-96fb-4766-b8d7-57cd4d199713\" (UID: \"91dd8c53-96fb-4766-b8d7-57cd4d199713\") " Dec 10 15:26:59 crc kubenswrapper[4847]: I1210 15:26:59.973968 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm" (OuterVolumeSpecName: "kube-api-access-2hlgm") pod "91dd8c53-96fb-4766-b8d7-57cd4d199713" (UID: "91dd8c53-96fb-4766-b8d7-57cd4d199713"). InnerVolumeSpecName "kube-api-access-2hlgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:27:00 crc kubenswrapper[4847]: I1210 15:27:00.075761 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hlgm\" (UniqueName: \"kubernetes.io/projected/91dd8c53-96fb-4766-b8d7-57cd4d199713-kube-api-access-2hlgm\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:00 crc kubenswrapper[4847]: I1210 15:27:00.120159 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "91dd8c53-96fb-4766-b8d7-57cd4d199713" (UID: "91dd8c53-96fb-4766-b8d7-57cd4d199713"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:27:00 crc kubenswrapper[4847]: I1210 15:27:00.179557 4847 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/91dd8c53-96fb-4766-b8d7-57cd4d199713-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:00 crc kubenswrapper[4847]: I1210 15:27:00.771670 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" path="/var/lib/kubelet/pods/91dd8c53-96fb-4766-b8d7-57cd4d199713/volumes" Dec 10 15:27:00 crc kubenswrapper[4847]: I1210 15:27:00.872761 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gk25n/must-gather-lvq8h" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.151001 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:03 crc kubenswrapper[4847]: E1210 15:27:03.152034 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="copy" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152052 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="copy" Dec 10 15:27:03 crc kubenswrapper[4847]: E1210 15:27:03.152074 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="registry-server" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152082 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="registry-server" Dec 10 15:27:03 crc kubenswrapper[4847]: E1210 15:27:03.152109 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="extract-content" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152134 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="extract-content" Dec 10 15:27:03 crc kubenswrapper[4847]: E1210 15:27:03.152152 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="gather" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152159 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="gather" Dec 10 15:27:03 crc kubenswrapper[4847]: E1210 15:27:03.152175 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="extract-utilities" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152183 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="extract-utilities" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152446 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="copy" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152467 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="91dd8c53-96fb-4766-b8d7-57cd4d199713" containerName="gather" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.152481 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d2085f-74d3-46c1-9994-56c64f04eed2" containerName="registry-server" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.154195 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.175089 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.236294 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svrl2\" (UniqueName: \"kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.236468 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.236529 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.338223 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svrl2\" (UniqueName: \"kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.338409 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.338482 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.339111 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.339163 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.359945 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svrl2\" (UniqueName: \"kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2\") pod \"certified-operators-flqpt\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:03 crc kubenswrapper[4847]: I1210 15:27:03.474928 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:04 crc kubenswrapper[4847]: I1210 15:27:04.032037 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:04 crc kubenswrapper[4847]: I1210 15:27:04.912810 4847 generic.go:334] "Generic (PLEG): container finished" podID="50aa0467-6b29-47bf-b543-256a747d33bc" containerID="5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de" exitCode=0 Dec 10 15:27:04 crc kubenswrapper[4847]: I1210 15:27:04.912865 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerDied","Data":"5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de"} Dec 10 15:27:04 crc kubenswrapper[4847]: I1210 15:27:04.913067 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerStarted","Data":"028cd4e9713f0000b70101e0ae3e457654016947b8af8c92ae99b931db327dcd"} Dec 10 15:27:04 crc kubenswrapper[4847]: I1210 15:27:04.918171 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.570761 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.573656 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.611212 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.680455 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27bdv\" (UniqueName: \"kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.680643 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.680672 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.782777 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27bdv\" (UniqueName: \"kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.782993 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.783043 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.784211 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.784589 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.802279 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27bdv\" (UniqueName: \"kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv\") pod \"redhat-marketplace-dkpj8\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:05 crc kubenswrapper[4847]: I1210 15:27:05.899751 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:06 crc kubenswrapper[4847]: I1210 15:27:06.296082 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:06 crc kubenswrapper[4847]: W1210 15:27:06.301795 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod643276e6_9228_416b_a33f_869733084691.slice/crio-0ee774d540c2b7d31aef88c5e7d30406ef53adf291b36535235a79aacfddd8bc WatchSource:0}: Error finding container 0ee774d540c2b7d31aef88c5e7d30406ef53adf291b36535235a79aacfddd8bc: Status 404 returned error can't find the container with id 0ee774d540c2b7d31aef88c5e7d30406ef53adf291b36535235a79aacfddd8bc Dec 10 15:27:06 crc kubenswrapper[4847]: I1210 15:27:06.933913 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerStarted","Data":"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92"} Dec 10 15:27:06 crc kubenswrapper[4847]: I1210 15:27:06.934138 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerStarted","Data":"0ee774d540c2b7d31aef88c5e7d30406ef53adf291b36535235a79aacfddd8bc"} Dec 10 15:27:07 crc kubenswrapper[4847]: I1210 15:27:07.945520 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerStarted","Data":"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff"} Dec 10 15:27:07 crc kubenswrapper[4847]: I1210 15:27:07.947783 4847 generic.go:334] "Generic (PLEG): container finished" podID="643276e6-9228-416b-a33f-869733084691" containerID="ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92" exitCode=0 Dec 10 15:27:07 crc kubenswrapper[4847]: I1210 15:27:07.947828 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerDied","Data":"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92"} Dec 10 15:27:07 crc kubenswrapper[4847]: I1210 15:27:07.947854 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerStarted","Data":"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02"} Dec 10 15:27:08 crc kubenswrapper[4847]: I1210 15:27:08.958785 4847 generic.go:334] "Generic (PLEG): container finished" podID="643276e6-9228-416b-a33f-869733084691" containerID="0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02" exitCode=0 Dec 10 15:27:08 crc kubenswrapper[4847]: I1210 15:27:08.958843 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerDied","Data":"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02"} Dec 10 15:27:08 crc kubenswrapper[4847]: I1210 15:27:08.961648 4847 generic.go:334] "Generic (PLEG): container finished" podID="50aa0467-6b29-47bf-b543-256a747d33bc" containerID="2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff" exitCode=0 Dec 10 15:27:08 crc kubenswrapper[4847]: I1210 15:27:08.961699 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerDied","Data":"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff"} Dec 10 15:27:09 crc kubenswrapper[4847]: I1210 15:27:09.759653 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:27:09 crc kubenswrapper[4847]: E1210 15:27:09.760039 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:27:09 crc kubenswrapper[4847]: I1210 15:27:09.976456 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerStarted","Data":"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b"} Dec 10 15:27:09 crc kubenswrapper[4847]: I1210 15:27:09.979597 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerStarted","Data":"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97"} Dec 10 15:27:10 crc kubenswrapper[4847]: I1210 15:27:10.001176 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-flqpt" podStartSLOduration=2.451033294 podStartE2EDuration="7.001154483s" podCreationTimestamp="2025-12-10 15:27:03 +0000 UTC" firstStartedPulling="2025-12-10 15:27:04.917810111 +0000 UTC m=+3774.487027751" lastFinishedPulling="2025-12-10 15:27:09.46793129 +0000 UTC m=+3779.037148940" observedRunningTime="2025-12-10 15:27:09.996420973 +0000 UTC m=+3779.565638623" watchObservedRunningTime="2025-12-10 15:27:10.001154483 +0000 UTC m=+3779.570372113" Dec 10 15:27:10 crc kubenswrapper[4847]: I1210 15:27:10.026120 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dkpj8" podStartSLOduration=2.591029383 podStartE2EDuration="5.026100343s" podCreationTimestamp="2025-12-10 15:27:05 +0000 UTC" firstStartedPulling="2025-12-10 15:27:06.994681699 +0000 UTC m=+3776.563899329" lastFinishedPulling="2025-12-10 15:27:09.429752649 +0000 UTC m=+3778.998970289" observedRunningTime="2025-12-10 15:27:10.018487695 +0000 UTC m=+3779.587705335" watchObservedRunningTime="2025-12-10 15:27:10.026100343 +0000 UTC m=+3779.595317973" Dec 10 15:27:13 crc kubenswrapper[4847]: I1210 15:27:13.475329 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:13 crc kubenswrapper[4847]: I1210 15:27:13.475798 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:13 crc kubenswrapper[4847]: I1210 15:27:13.545183 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:14 crc kubenswrapper[4847]: I1210 15:27:14.056290 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:14 crc kubenswrapper[4847]: I1210 15:27:14.741966 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:15 crc kubenswrapper[4847]: I1210 15:27:15.900536 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:15 crc kubenswrapper[4847]: I1210 15:27:15.900591 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:15 crc kubenswrapper[4847]: I1210 15:27:15.951806 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.029684 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-flqpt" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="registry-server" containerID="cri-o://29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b" gracePeriod=2 Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.088173 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.479678 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.637441 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svrl2\" (UniqueName: \"kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2\") pod \"50aa0467-6b29-47bf-b543-256a747d33bc\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.637639 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content\") pod \"50aa0467-6b29-47bf-b543-256a747d33bc\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.637685 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities\") pod \"50aa0467-6b29-47bf-b543-256a747d33bc\" (UID: \"50aa0467-6b29-47bf-b543-256a747d33bc\") " Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.638796 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities" (OuterVolumeSpecName: "utilities") pod "50aa0467-6b29-47bf-b543-256a747d33bc" (UID: "50aa0467-6b29-47bf-b543-256a747d33bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.644670 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2" (OuterVolumeSpecName: "kube-api-access-svrl2") pod "50aa0467-6b29-47bf-b543-256a747d33bc" (UID: "50aa0467-6b29-47bf-b543-256a747d33bc"). InnerVolumeSpecName "kube-api-access-svrl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.700412 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50aa0467-6b29-47bf-b543-256a747d33bc" (UID: "50aa0467-6b29-47bf-b543-256a747d33bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.739796 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.740141 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50aa0467-6b29-47bf-b543-256a747d33bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:16 crc kubenswrapper[4847]: I1210 15:27:16.740152 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svrl2\" (UniqueName: \"kubernetes.io/projected/50aa0467-6b29-47bf-b543-256a747d33bc-kube-api-access-svrl2\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.044265 4847 generic.go:334] "Generic (PLEG): container finished" podID="50aa0467-6b29-47bf-b543-256a747d33bc" containerID="29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b" exitCode=0 Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.044326 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flqpt" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.044321 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerDied","Data":"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b"} Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.044379 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flqpt" event={"ID":"50aa0467-6b29-47bf-b543-256a747d33bc","Type":"ContainerDied","Data":"028cd4e9713f0000b70101e0ae3e457654016947b8af8c92ae99b931db327dcd"} Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.044410 4847 scope.go:117] "RemoveContainer" containerID="29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.079914 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.085988 4847 scope.go:117] "RemoveContainer" containerID="2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.088768 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-flqpt"] Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.110921 4847 scope.go:117] "RemoveContainer" containerID="5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.149254 4847 scope.go:117] "RemoveContainer" containerID="29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b" Dec 10 15:27:17 crc kubenswrapper[4847]: E1210 15:27:17.149797 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b\": container with ID starting with 29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b not found: ID does not exist" containerID="29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.149862 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b"} err="failed to get container status \"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b\": rpc error: code = NotFound desc = could not find container \"29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b\": container with ID starting with 29c68ebc03297935459ac9977cf4550948ed52ed52306f1723c67816c5d6b99b not found: ID does not exist" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.149896 4847 scope.go:117] "RemoveContainer" containerID="2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff" Dec 10 15:27:17 crc kubenswrapper[4847]: E1210 15:27:17.150223 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff\": container with ID starting with 2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff not found: ID does not exist" containerID="2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.150255 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff"} err="failed to get container status \"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff\": rpc error: code = NotFound desc = could not find container \"2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff\": container with ID starting with 2020332a976a06a0dcfdb16f793d506edc15d4639b23b0c340a907f2e3da09ff not found: ID does not exist" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.150276 4847 scope.go:117] "RemoveContainer" containerID="5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de" Dec 10 15:27:17 crc kubenswrapper[4847]: E1210 15:27:17.150684 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de\": container with ID starting with 5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de not found: ID does not exist" containerID="5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de" Dec 10 15:27:17 crc kubenswrapper[4847]: I1210 15:27:17.150738 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de"} err="failed to get container status \"5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de\": rpc error: code = NotFound desc = could not find container \"5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de\": container with ID starting with 5aa896073164f0b45fffbffb4ae30ad7f2cff1a770238607ec1b12ef54c153de not found: ID does not exist" Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.352623 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.353680 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dkpj8" podUID="643276e6-9228-416b-a33f-869733084691" containerName="registry-server" containerID="cri-o://a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97" gracePeriod=2 Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.770808 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" path="/var/lib/kubelet/pods/50aa0467-6b29-47bf-b543-256a747d33bc/volumes" Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.838861 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.980800 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities\") pod \"643276e6-9228-416b-a33f-869733084691\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.981364 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content\") pod \"643276e6-9228-416b-a33f-869733084691\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.981623 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27bdv\" (UniqueName: \"kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv\") pod \"643276e6-9228-416b-a33f-869733084691\" (UID: \"643276e6-9228-416b-a33f-869733084691\") " Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.981940 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities" (OuterVolumeSpecName: "utilities") pod "643276e6-9228-416b-a33f-869733084691" (UID: "643276e6-9228-416b-a33f-869733084691"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.982684 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:18 crc kubenswrapper[4847]: I1210 15:27:18.992165 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv" (OuterVolumeSpecName: "kube-api-access-27bdv") pod "643276e6-9228-416b-a33f-869733084691" (UID: "643276e6-9228-416b-a33f-869733084691"). InnerVolumeSpecName "kube-api-access-27bdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.001687 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "643276e6-9228-416b-a33f-869733084691" (UID: "643276e6-9228-416b-a33f-869733084691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.067665 4847 generic.go:334] "Generic (PLEG): container finished" podID="643276e6-9228-416b-a33f-869733084691" containerID="a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97" exitCode=0 Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.067764 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerDied","Data":"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97"} Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.067823 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkpj8" event={"ID":"643276e6-9228-416b-a33f-869733084691","Type":"ContainerDied","Data":"0ee774d540c2b7d31aef88c5e7d30406ef53adf291b36535235a79aacfddd8bc"} Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.067847 4847 scope.go:117] "RemoveContainer" containerID="a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.068132 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkpj8" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.085237 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27bdv\" (UniqueName: \"kubernetes.io/projected/643276e6-9228-416b-a33f-869733084691-kube-api-access-27bdv\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.085268 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/643276e6-9228-416b-a33f-869733084691-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.092543 4847 scope.go:117] "RemoveContainer" containerID="0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.111878 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.123662 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkpj8"] Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.125551 4847 scope.go:117] "RemoveContainer" containerID="ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.169553 4847 scope.go:117] "RemoveContainer" containerID="a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97" Dec 10 15:27:19 crc kubenswrapper[4847]: E1210 15:27:19.170058 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97\": container with ID starting with a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97 not found: ID does not exist" containerID="a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.170215 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97"} err="failed to get container status \"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97\": rpc error: code = NotFound desc = could not find container \"a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97\": container with ID starting with a57db5ee87b8d15d22753a9dac43d9fd934632d3e5f40116e6e9901990a74f97 not found: ID does not exist" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.170393 4847 scope.go:117] "RemoveContainer" containerID="0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02" Dec 10 15:27:19 crc kubenswrapper[4847]: E1210 15:27:19.171393 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02\": container with ID starting with 0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02 not found: ID does not exist" containerID="0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.171451 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02"} err="failed to get container status \"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02\": rpc error: code = NotFound desc = could not find container \"0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02\": container with ID starting with 0054f6c3a1a3e8b7db099d6c4a5506bcbc09ec090b426cf0ea43cc609fb58e02 not found: ID does not exist" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.171489 4847 scope.go:117] "RemoveContainer" containerID="ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92" Dec 10 15:27:19 crc kubenswrapper[4847]: E1210 15:27:19.171893 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92\": container with ID starting with ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92 not found: ID does not exist" containerID="ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92" Dec 10 15:27:19 crc kubenswrapper[4847]: I1210 15:27:19.171939 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92"} err="failed to get container status \"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92\": rpc error: code = NotFound desc = could not find container \"ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92\": container with ID starting with ed230f557d40016c0fe02e8ae96569aae9350edb7d42b085dfefb6377edaae92 not found: ID does not exist" Dec 10 15:27:20 crc kubenswrapper[4847]: I1210 15:27:20.770800 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="643276e6-9228-416b-a33f-869733084691" path="/var/lib/kubelet/pods/643276e6-9228-416b-a33f-869733084691/volumes" Dec 10 15:27:22 crc kubenswrapper[4847]: I1210 15:27:22.760389 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:27:22 crc kubenswrapper[4847]: E1210 15:27:22.761120 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:27:35 crc kubenswrapper[4847]: I1210 15:27:35.759752 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:27:35 crc kubenswrapper[4847]: E1210 15:27:35.762359 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:27:47 crc kubenswrapper[4847]: I1210 15:27:47.760427 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:27:47 crc kubenswrapper[4847]: E1210 15:27:47.761532 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:00 crc kubenswrapper[4847]: I1210 15:28:00.766023 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:00 crc kubenswrapper[4847]: E1210 15:28:00.766884 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:12 crc kubenswrapper[4847]: I1210 15:28:12.759765 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:12 crc kubenswrapper[4847]: E1210 15:28:12.760693 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:23 crc kubenswrapper[4847]: I1210 15:28:23.760620 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:23 crc kubenswrapper[4847]: E1210 15:28:23.761470 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:25 crc kubenswrapper[4847]: I1210 15:28:25.105164 4847 scope.go:117] "RemoveContainer" containerID="3497ea0f03db7cb0d9f27feba5823de6e578e662bbee3d907ebf199aa18c467c" Dec 10 15:28:25 crc kubenswrapper[4847]: I1210 15:28:25.218240 4847 scope.go:117] "RemoveContainer" containerID="8c1a4ece23b596436592e9c879b9e4922967b996c1d57100bf27d060b4813c38" Dec 10 15:28:25 crc kubenswrapper[4847]: I1210 15:28:25.237572 4847 scope.go:117] "RemoveContainer" containerID="662af327dc13a4b8333810aed051d1e59b31fbe26c156817922fb9179a59012e" Dec 10 15:28:35 crc kubenswrapper[4847]: I1210 15:28:35.761633 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:35 crc kubenswrapper[4847]: E1210 15:28:35.762407 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:46 crc kubenswrapper[4847]: I1210 15:28:46.760884 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:46 crc kubenswrapper[4847]: E1210 15:28:46.762032 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:28:57 crc kubenswrapper[4847]: I1210 15:28:57.760105 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:28:57 crc kubenswrapper[4847]: E1210 15:28:57.761224 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:29:11 crc kubenswrapper[4847]: I1210 15:29:11.759438 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:29:12 crc kubenswrapper[4847]: I1210 15:29:12.322246 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858"} Dec 10 15:29:25 crc kubenswrapper[4847]: I1210 15:29:25.322440 4847 scope.go:117] "RemoveContainer" containerID="b8e8ad41c551048b75d5bafdf7489dec80783bf9e55cc85c7d931b3d19ee4c64" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.188784 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt"] Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.189914 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643276e6-9228-416b-a33f-869733084691" containerName="extract-content" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.189934 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="643276e6-9228-416b-a33f-869733084691" containerName="extract-content" Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.189956 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643276e6-9228-416b-a33f-869733084691" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.189963 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="643276e6-9228-416b-a33f-869733084691" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.189984 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643276e6-9228-416b-a33f-869733084691" containerName="extract-utilities" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.189993 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="643276e6-9228-416b-a33f-869733084691" containerName="extract-utilities" Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.190011 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="extract-utilities" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.190017 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="extract-utilities" Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.190034 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.190041 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: E1210 15:30:00.190059 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="extract-content" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.190068 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="extract-content" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.190274 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="50aa0467-6b29-47bf-b543-256a747d33bc" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.190309 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="643276e6-9228-416b-a33f-869733084691" containerName="registry-server" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.191127 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.193416 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.193857 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.202708 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt"] Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.316011 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.316077 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtmvm\" (UniqueName: \"kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.316099 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.418103 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.418167 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtmvm\" (UniqueName: \"kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.418190 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.419505 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.426420 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.442574 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtmvm\" (UniqueName: \"kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm\") pod \"collect-profiles-29423010-6jvtt\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.515876 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:00 crc kubenswrapper[4847]: I1210 15:30:00.991430 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt"] Dec 10 15:30:01 crc kubenswrapper[4847]: I1210 15:30:01.803800 4847 generic.go:334] "Generic (PLEG): container finished" podID="2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" containerID="570f4405e94c6c2a987cb902292c9dcb1477c0d816209120173772c6e18975f8" exitCode=0 Dec 10 15:30:01 crc kubenswrapper[4847]: I1210 15:30:01.803998 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" event={"ID":"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84","Type":"ContainerDied","Data":"570f4405e94c6c2a987cb902292c9dcb1477c0d816209120173772c6e18975f8"} Dec 10 15:30:01 crc kubenswrapper[4847]: I1210 15:30:01.804231 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" event={"ID":"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84","Type":"ContainerStarted","Data":"9354fdb44ac390ea672bb2b9fc1f8be051f8e7ccc91249b67bff10fcb88ae514"} Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.601184 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jsgq/must-gather-q76lt"] Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.603374 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.606337 4847 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5jsgq"/"default-dockercfg-tvgf8" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.606798 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5jsgq"/"openshift-service-ca.crt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.609588 4847 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5jsgq"/"kube-root-ca.crt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.612829 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5jsgq/must-gather-q76lt"] Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.762996 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c76kt\" (UniqueName: \"kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.763411 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.864715 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.864849 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c76kt\" (UniqueName: \"kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.866273 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.908566 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c76kt\" (UniqueName: \"kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt\") pod \"must-gather-q76lt\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:02 crc kubenswrapper[4847]: I1210 15:30:02.929980 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.273701 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.373216 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume\") pod \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.373318 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume\") pod \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.373395 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtmvm\" (UniqueName: \"kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm\") pod \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\" (UID: \"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84\") " Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.374496 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" (UID: "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.379247 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm" (OuterVolumeSpecName: "kube-api-access-qtmvm") pod "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" (UID: "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84"). InnerVolumeSpecName "kube-api-access-qtmvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.379383 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" (UID: "2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.439594 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5jsgq/must-gather-q76lt"] Dec 10 15:30:03 crc kubenswrapper[4847]: W1210 15:30:03.443336 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5233e69a_8779_4a5a_945f_3acf0bcf4334.slice/crio-dceb190421a46a2225a30f268eb87c96e6a08aa6b5965baa77ef801fa80edfa1 WatchSource:0}: Error finding container dceb190421a46a2225a30f268eb87c96e6a08aa6b5965baa77ef801fa80edfa1: Status 404 returned error can't find the container with id dceb190421a46a2225a30f268eb87c96e6a08aa6b5965baa77ef801fa80edfa1 Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.475878 4847 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.476059 4847 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.476157 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtmvm\" (UniqueName: \"kubernetes.io/projected/2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84-kube-api-access-qtmvm\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.866001 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/must-gather-q76lt" event={"ID":"5233e69a-8779-4a5a-945f-3acf0bcf4334","Type":"ContainerStarted","Data":"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3"} Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.866071 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/must-gather-q76lt" event={"ID":"5233e69a-8779-4a5a-945f-3acf0bcf4334","Type":"ContainerStarted","Data":"dceb190421a46a2225a30f268eb87c96e6a08aa6b5965baa77ef801fa80edfa1"} Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.867815 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" event={"ID":"2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84","Type":"ContainerDied","Data":"9354fdb44ac390ea672bb2b9fc1f8be051f8e7ccc91249b67bff10fcb88ae514"} Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.867854 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423010-6jvtt" Dec 10 15:30:03 crc kubenswrapper[4847]: I1210 15:30:03.867864 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9354fdb44ac390ea672bb2b9fc1f8be051f8e7ccc91249b67bff10fcb88ae514" Dec 10 15:30:04 crc kubenswrapper[4847]: I1210 15:30:04.382495 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7"] Dec 10 15:30:04 crc kubenswrapper[4847]: I1210 15:30:04.391525 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29422965-pptl7"] Dec 10 15:30:04 crc kubenswrapper[4847]: I1210 15:30:04.771081 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="183c9599-f0db-4e64-9ad1-3c4b60bef02c" path="/var/lib/kubelet/pods/183c9599-f0db-4e64-9ad1-3c4b60bef02c/volumes" Dec 10 15:30:04 crc kubenswrapper[4847]: I1210 15:30:04.878337 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/must-gather-q76lt" event={"ID":"5233e69a-8779-4a5a-945f-3acf0bcf4334","Type":"ContainerStarted","Data":"ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6"} Dec 10 15:30:04 crc kubenswrapper[4847]: I1210 15:30:04.901606 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5jsgq/must-gather-q76lt" podStartSLOduration=2.901587347 podStartE2EDuration="2.901587347s" podCreationTimestamp="2025-12-10 15:30:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 15:30:04.89580377 +0000 UTC m=+3954.465021400" watchObservedRunningTime="2025-12-10 15:30:04.901587347 +0000 UTC m=+3954.470804977" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.659037 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-77rfg"] Dec 10 15:30:07 crc kubenswrapper[4847]: E1210 15:30:07.659900 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" containerName="collect-profiles" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.659913 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" containerName="collect-profiles" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.660099 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2fbe3f-f39c-4cf0-9698-e6fefd2b1a84" containerName="collect-profiles" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.660705 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.773929 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn74c\" (UniqueName: \"kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.774007 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.876433 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn74c\" (UniqueName: \"kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.876899 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.877919 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.907328 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn74c\" (UniqueName: \"kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c\") pod \"crc-debug-77rfg\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:07 crc kubenswrapper[4847]: I1210 15:30:07.981142 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:08 crc kubenswrapper[4847]: W1210 15:30:08.013606 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1e7b0f_9a6f_44a0_b058_db76f9bff9dc.slice/crio-ef2d55a0f3a97bcd7549f7b0d88a85f1bea360a21aaf7c83c33627bbed0189e7 WatchSource:0}: Error finding container ef2d55a0f3a97bcd7549f7b0d88a85f1bea360a21aaf7c83c33627bbed0189e7: Status 404 returned error can't find the container with id ef2d55a0f3a97bcd7549f7b0d88a85f1bea360a21aaf7c83c33627bbed0189e7 Dec 10 15:30:08 crc kubenswrapper[4847]: I1210 15:30:08.922324 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" event={"ID":"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc","Type":"ContainerStarted","Data":"35f176381fdd531eb5d71f21a07b5264f780bfb4bd53a7bca353ec0890ce8f06"} Dec 10 15:30:08 crc kubenswrapper[4847]: I1210 15:30:08.922966 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" event={"ID":"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc","Type":"ContainerStarted","Data":"ef2d55a0f3a97bcd7549f7b0d88a85f1bea360a21aaf7c83c33627bbed0189e7"} Dec 10 15:30:08 crc kubenswrapper[4847]: I1210 15:30:08.954710 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" podStartSLOduration=1.954674021 podStartE2EDuration="1.954674021s" podCreationTimestamp="2025-12-10 15:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 15:30:08.935425811 +0000 UTC m=+3958.504643451" watchObservedRunningTime="2025-12-10 15:30:08.954674021 +0000 UTC m=+3958.523891651" Dec 10 15:30:25 crc kubenswrapper[4847]: I1210 15:30:25.379850 4847 scope.go:117] "RemoveContainer" containerID="4b74eee694f848bebd50249943b03c3be4761073a25469c692597515fc7d0a3f" Dec 10 15:30:44 crc kubenswrapper[4847]: I1210 15:30:44.225446 4847 generic.go:334] "Generic (PLEG): container finished" podID="9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" containerID="35f176381fdd531eb5d71f21a07b5264f780bfb4bd53a7bca353ec0890ce8f06" exitCode=0 Dec 10 15:30:44 crc kubenswrapper[4847]: I1210 15:30:44.225523 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" event={"ID":"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc","Type":"ContainerDied","Data":"35f176381fdd531eb5d71f21a07b5264f780bfb4bd53a7bca353ec0890ce8f06"} Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.343417 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.382874 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-77rfg"] Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.392564 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-77rfg"] Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.414245 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn74c\" (UniqueName: \"kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c\") pod \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.414424 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host\") pod \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\" (UID: \"9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc\") " Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.415078 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host" (OuterVolumeSpecName: "host") pod "9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" (UID: "9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.421614 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c" (OuterVolumeSpecName: "kube-api-access-jn74c") pod "9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" (UID: "9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc"). InnerVolumeSpecName "kube-api-access-jn74c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.516318 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn74c\" (UniqueName: \"kubernetes.io/projected/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-kube-api-access-jn74c\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:45 crc kubenswrapper[4847]: I1210 15:30:45.516358 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.246098 4847 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef2d55a0f3a97bcd7549f7b0d88a85f1bea360a21aaf7c83c33627bbed0189e7" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.246199 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-77rfg" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.555518 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-k256n"] Dec 10 15:30:46 crc kubenswrapper[4847]: E1210 15:30:46.557005 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" containerName="container-00" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.557165 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" containerName="container-00" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.557594 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" containerName="container-00" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.558755 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.636270 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8c8\" (UniqueName: \"kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.636479 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.738315 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.738659 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8c8\" (UniqueName: \"kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.738490 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.757981 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8c8\" (UniqueName: \"kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8\") pod \"crc-debug-k256n\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.773027 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc" path="/var/lib/kubelet/pods/9c1e7b0f-9a6f-44a0-b058-db76f9bff9dc/volumes" Dec 10 15:30:46 crc kubenswrapper[4847]: I1210 15:30:46.879622 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:47 crc kubenswrapper[4847]: I1210 15:30:47.264353 4847 generic.go:334] "Generic (PLEG): container finished" podID="550152e3-6a62-4b95-b9d3-4b72fabcc1bc" containerID="08d5443edc06b4a59416cdf4bb05930cc8dc635942e7e311159d2a6d7568b346" exitCode=0 Dec 10 15:30:47 crc kubenswrapper[4847]: I1210 15:30:47.264766 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-k256n" event={"ID":"550152e3-6a62-4b95-b9d3-4b72fabcc1bc","Type":"ContainerDied","Data":"08d5443edc06b4a59416cdf4bb05930cc8dc635942e7e311159d2a6d7568b346"} Dec 10 15:30:47 crc kubenswrapper[4847]: I1210 15:30:47.265130 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-k256n" event={"ID":"550152e3-6a62-4b95-b9d3-4b72fabcc1bc","Type":"ContainerStarted","Data":"bf0bbc0148a3b814813cea3002ef991c449657683d694a19c2e776ae39745928"} Dec 10 15:30:47 crc kubenswrapper[4847]: I1210 15:30:47.813285 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-k256n"] Dec 10 15:30:47 crc kubenswrapper[4847]: I1210 15:30:47.822406 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-k256n"] Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.486567 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.574505 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host\") pod \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.574640 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host" (OuterVolumeSpecName: "host") pod "550152e3-6a62-4b95-b9d3-4b72fabcc1bc" (UID: "550152e3-6a62-4b95-b9d3-4b72fabcc1bc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.574700 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn8c8\" (UniqueName: \"kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8\") pod \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\" (UID: \"550152e3-6a62-4b95-b9d3-4b72fabcc1bc\") " Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.575161 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.584994 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8" (OuterVolumeSpecName: "kube-api-access-dn8c8") pod "550152e3-6a62-4b95-b9d3-4b72fabcc1bc" (UID: "550152e3-6a62-4b95-b9d3-4b72fabcc1bc"). InnerVolumeSpecName "kube-api-access-dn8c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.676629 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn8c8\" (UniqueName: \"kubernetes.io/projected/550152e3-6a62-4b95-b9d3-4b72fabcc1bc-kube-api-access-dn8c8\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:48 crc kubenswrapper[4847]: I1210 15:30:48.769394 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="550152e3-6a62-4b95-b9d3-4b72fabcc1bc" path="/var/lib/kubelet/pods/550152e3-6a62-4b95-b9d3-4b72fabcc1bc/volumes" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.284139 4847 scope.go:117] "RemoveContainer" containerID="08d5443edc06b4a59416cdf4bb05930cc8dc635942e7e311159d2a6d7568b346" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.284171 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-k256n" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.481618 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-tdjc6"] Dec 10 15:30:49 crc kubenswrapper[4847]: E1210 15:30:49.482500 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="550152e3-6a62-4b95-b9d3-4b72fabcc1bc" containerName="container-00" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.482519 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="550152e3-6a62-4b95-b9d3-4b72fabcc1bc" containerName="container-00" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.482693 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="550152e3-6a62-4b95-b9d3-4b72fabcc1bc" containerName="container-00" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.483312 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.592863 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.593006 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ztvb\" (UniqueName: \"kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.695227 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.695304 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ztvb\" (UniqueName: \"kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:49 crc kubenswrapper[4847]: I1210 15:30:49.695459 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:50 crc kubenswrapper[4847]: I1210 15:30:50.097626 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ztvb\" (UniqueName: \"kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb\") pod \"crc-debug-tdjc6\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:50 crc kubenswrapper[4847]: I1210 15:30:50.104174 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:51 crc kubenswrapper[4847]: I1210 15:30:51.321581 4847 generic.go:334] "Generic (PLEG): container finished" podID="81c83b57-682f-41cd-9204-d1316a20f8f4" containerID="c8d351c8142cef124015865500264c95c94b153c68d627fd4a5d98083df0cf0d" exitCode=0 Dec 10 15:30:51 crc kubenswrapper[4847]: I1210 15:30:51.321701 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" event={"ID":"81c83b57-682f-41cd-9204-d1316a20f8f4","Type":"ContainerDied","Data":"c8d351c8142cef124015865500264c95c94b153c68d627fd4a5d98083df0cf0d"} Dec 10 15:30:51 crc kubenswrapper[4847]: I1210 15:30:51.321936 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" event={"ID":"81c83b57-682f-41cd-9204-d1316a20f8f4","Type":"ContainerStarted","Data":"913d586a4b3663621c4ddfe5f2faa5b49e975dccf04c9a9f1ebe0db50944f170"} Dec 10 15:30:51 crc kubenswrapper[4847]: I1210 15:30:51.382818 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-tdjc6"] Dec 10 15:30:51 crc kubenswrapper[4847]: I1210 15:30:51.391372 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jsgq/crc-debug-tdjc6"] Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.433150 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.551832 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host\") pod \"81c83b57-682f-41cd-9204-d1316a20f8f4\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.552017 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ztvb\" (UniqueName: \"kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb\") pod \"81c83b57-682f-41cd-9204-d1316a20f8f4\" (UID: \"81c83b57-682f-41cd-9204-d1316a20f8f4\") " Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.552999 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host" (OuterVolumeSpecName: "host") pod "81c83b57-682f-41cd-9204-d1316a20f8f4" (UID: "81c83b57-682f-41cd-9204-d1316a20f8f4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.558537 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb" (OuterVolumeSpecName: "kube-api-access-2ztvb") pod "81c83b57-682f-41cd-9204-d1316a20f8f4" (UID: "81c83b57-682f-41cd-9204-d1316a20f8f4"). InnerVolumeSpecName "kube-api-access-2ztvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.654054 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ztvb\" (UniqueName: \"kubernetes.io/projected/81c83b57-682f-41cd-9204-d1316a20f8f4-kube-api-access-2ztvb\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.654094 4847 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/81c83b57-682f-41cd-9204-d1316a20f8f4-host\") on node \"crc\" DevicePath \"\"" Dec 10 15:30:52 crc kubenswrapper[4847]: I1210 15:30:52.775621 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81c83b57-682f-41cd-9204-d1316a20f8f4" path="/var/lib/kubelet/pods/81c83b57-682f-41cd-9204-d1316a20f8f4/volumes" Dec 10 15:30:53 crc kubenswrapper[4847]: I1210 15:30:53.341470 4847 scope.go:117] "RemoveContainer" containerID="c8d351c8142cef124015865500264c95c94b153c68d627fd4a5d98083df0cf0d" Dec 10 15:30:53 crc kubenswrapper[4847]: I1210 15:30:53.341640 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/crc-debug-tdjc6" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.359575 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5488ccdfb4-gmxkx_e4d02170-155d-4380-9e77-e8705b47a968/barbican-api/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.441627 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5488ccdfb4-gmxkx_e4d02170-155d-4380-9e77-e8705b47a968/barbican-api-log/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.585674 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9fcb8ccdb-j6kdw_70fb729f-c877-4d76-a5b0-a84470c6c4bc/barbican-keystone-listener-log/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.603054 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9fcb8ccdb-j6kdw_70fb729f-c877-4d76-a5b0-a84470c6c4bc/barbican-keystone-listener/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.666944 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b85c87b55-fkkkc_9e5b990d-5058-4827-846b-e5ba776fa0ba/barbican-worker/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.777256 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-b85c87b55-fkkkc_9e5b990d-5058-4827-846b-e5ba776fa0ba/barbican-worker-log/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.900537 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-v2kpm_c65ca249-2dbf-48c4-87c5-9ff2d9bd72b6/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:15 crc kubenswrapper[4847]: I1210 15:31:15.999477 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/ceilometer-central-agent/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.042343 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/ceilometer-notification-agent/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.092149 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/proxy-httpd/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.153890 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3cd44098-3d49-408d-8478-fceaa9ac7e57/sg-core/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.342153 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f0c3090a-89da-46f1-9998-a68188814b05/cinder-api/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.354286 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f0c3090a-89da-46f1-9998-a68188814b05/cinder-api-log/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.531912 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6b8b833f-7fa3-4232-831a-18d4ba96304b/cinder-scheduler/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.601458 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6b8b833f-7fa3-4232-831a-18d4ba96304b/probe/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.675363 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-fh2mf_21486092-46dc-4bf8-9cfc-5fc5781461c2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:16 crc kubenswrapper[4847]: I1210 15:31:16.856087 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hcfd7_da983cff-081c-4811-9ed3-f3cc109489bd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.107396 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/init/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.306674 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/dnsmasq-dns/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.338501 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-mbxst_70f2cfd7-1fd9-48f5-8004-6df8f4d428c9/init/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.367988 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-krj6h_48c5aa38-4902-4105-99ac-40fb8ab5a288/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.501028 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23c6ec92-8230-4322-ad9c-67c0a3260eec/glance-httpd/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.535274 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_23c6ec92-8230-4322-ad9c-67c0a3260eec/glance-log/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.721412 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4454a179-0fc2-4623-b1ba-b7946009d769/glance-httpd/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.725146 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_4454a179-0fc2-4623-b1ba-b7946009d769/glance-log/0.log" Dec 10 15:31:17 crc kubenswrapper[4847]: I1210 15:31:17.895682 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4f6b7576-wvhs4_f7e7178d-cd26-4623-a05c-6eee95cb5dd6/horizon/0.log" Dec 10 15:31:18 crc kubenswrapper[4847]: I1210 15:31:18.650930 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jlctv_a1a6fb9c-79e5-4603-8849-b3f36f0a6ec1/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:18 crc kubenswrapper[4847]: I1210 15:31:18.879220 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4f6b7576-wvhs4_f7e7178d-cd26-4623-a05c-6eee95cb5dd6/horizon-log/0.log" Dec 10 15:31:18 crc kubenswrapper[4847]: I1210 15:31:18.913632 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-hvd6g_d4c9d846-3adb-4f21-8945-f27d14123b1e/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.114672 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54599d665c-kl29l_8f2f4d1f-1fc9-48cd-8697-bdc788483a1c/keystone-api/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.133935 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29422981-q652q_b0790548-d5d8-4688-9120-bb942f395fd2/keystone-cron/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.277873 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6511a21a-53cd-4ecc-a550-cc4a3ef72c8c/kube-state-metrics/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.357977 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4n4rp_eb5a7f57-0eb5-47f9-ae7c-173f849c7a02/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.773735 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c86c8879c-drvf5_7d85a520-fbc1-4a46-b33e-7513646bfeaf/neutron-api/0.log" Dec 10 15:31:19 crc kubenswrapper[4847]: I1210 15:31:19.781101 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c86c8879c-drvf5_7d85a520-fbc1-4a46-b33e-7513646bfeaf/neutron-httpd/0.log" Dec 10 15:31:20 crc kubenswrapper[4847]: I1210 15:31:20.231166 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-b442c_04d0cfc8-2361-4e74-bacc-379a98054d97/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:20 crc kubenswrapper[4847]: I1210 15:31:20.806900 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0205d3f1-fa73-4291-a22b-1bdd4aaa23ec/nova-api-log/0.log" Dec 10 15:31:20 crc kubenswrapper[4847]: I1210 15:31:20.838093 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e9315ad6-b2ce-420a-b8a4-c168264662de/nova-cell0-conductor-conductor/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.248658 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0205d3f1-fa73-4291-a22b-1bdd4aaa23ec/nova-api-api/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.253765 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1598cbc7-d655-494f-a2b1-ddbd4db56aa5/nova-cell1-conductor-conductor/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.401657 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_d1d12a43-2aba-411a-ad3d-628d6f468c95/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.496864 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ct6mn_9b520035-682d-43ae-acba-403455d78c9f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.686442 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d7f8698b-dd55-4811-81da-55dac46eeffe/nova-metadata-log/0.log" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.806184 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:21 crc kubenswrapper[4847]: E1210 15:31:21.806649 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81c83b57-682f-41cd-9204-d1316a20f8f4" containerName="container-00" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.806667 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="81c83b57-682f-41cd-9204-d1316a20f8f4" containerName="container-00" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.806887 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="81c83b57-682f-41cd-9204-d1316a20f8f4" containerName="container-00" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.808388 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.834925 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.869033 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.869460 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.869552 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6bkg\" (UniqueName: \"kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.971447 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.971557 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.971632 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6bkg\" (UniqueName: \"kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.972173 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.972222 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:21 crc kubenswrapper[4847]: I1210 15:31:21.996614 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6bkg\" (UniqueName: \"kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg\") pod \"redhat-operators-qcncf\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.037901 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/mysql-bootstrap/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.072420 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8b228613-6a71-4c3b-a95a-e0a496c0b346/nova-scheduler-scheduler/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.162032 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.354695 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/galera/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.375660 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4499b911-76a6-4ceb-8320-c87af3f1bd0a/mysql-bootstrap/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.596283 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/mysql-bootstrap/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.712548 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.982529 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/mysql-bootstrap/0.log" Dec 10 15:31:22 crc kubenswrapper[4847]: I1210 15:31:22.997950 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_32d8cf5e-b11b-469d-8838-b27e201d95c0/galera/0.log" Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.246879 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2559a413-0d27-43f0-8fa3-2da6e9d66eff/openstackclient/0.log" Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.267107 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2xjh7_b962e42f-e21b-4016-9587-111fb6eb68ac/ovn-controller/0.log" Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.510073 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d7f8698b-dd55-4811-81da-55dac46eeffe/nova-metadata-metadata/0.log" Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.619252 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-n4rbc_4f080a9d-d07e-4a5e-b782-245d4c7b31cf/openstack-network-exporter/0.log" Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.669249 4847 generic.go:334] "Generic (PLEG): container finished" podID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerID="038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b" exitCode=0 Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.669305 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerDied","Data":"038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b"} Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.669341 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerStarted","Data":"bee3d4ce7c606106898f50915900a41a47a035a650210f08d556195a30b58964"} Dec 10 15:31:23 crc kubenswrapper[4847]: I1210 15:31:23.818585 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server-init/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.029123 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovs-vswitchd/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.070976 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server-init/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.080891 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rnkkd_6f0ee780-f899-462e-89dd-e353a5790288/ovsdb-server/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.289620 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vsngf_dfed455d-08ba-417a-975b-09dc84eaec54/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.316844 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a4d0db06-c1bd-4af6-b78e-a1703d914c6d/openstack-network-exporter/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.408055 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a4d0db06-c1bd-4af6-b78e-a1703d914c6d/ovn-northd/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.557707 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_26654ec0-c7fa-44da-9be9-8d5218addac5/openstack-network-exporter/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.614169 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_26654ec0-c7fa-44da-9be9-8d5218addac5/ovsdbserver-nb/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.681647 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerStarted","Data":"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09"} Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.746076 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_beaec882-8047-4348-bcf5-a45eed4dba20/openstack-network-exporter/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.799247 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_beaec882-8047-4348-bcf5-a45eed4dba20/ovsdbserver-sb/0.log" Dec 10 15:31:24 crc kubenswrapper[4847]: I1210 15:31:24.983093 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d896bd648-rk9qk_7895251f-c0bb-40b7-a4a2-802a3517cd98/placement-api/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.156642 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6d896bd648-rk9qk_7895251f-c0bb-40b7-a4a2-802a3517cd98/placement-log/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.287213 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/setup-container/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.611975 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/setup-container/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.638914 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/setup-container/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.651512 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0cca7d7b-5f9b-4d29-84c2-bccda46b9ebf/rabbitmq/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.883663 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/setup-container/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.904380 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zghlr_4ef271cb-44e6-4f26-95fc-3ca47f6ab290/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:25 crc kubenswrapper[4847]: I1210 15:31:25.946496 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_75428316-3c6c-4166-a55e-cb293dbad536/rabbitmq/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.180361 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9j7sd_1f79a99b-09b0-4951-bf05-42fe68d44e53/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.210216 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s87nz_d39485c7-eaab-4829-b978-ca8109f82227/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.476518 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-xs56g_f9991c0e-7804-406a-bc24-92c5d5d7fa61/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.489770 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-4ckkd_6e914d68-6bc2-420a-b0cb-3f424d71652c/ssh-known-hosts-edpm-deployment/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.788387 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65d99c95b7-t6kg4_9cd88676-59ca-4c55-aaf7-caab8965d2bb/proxy-server/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.844483 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65d99c95b7-t6kg4_9cd88676-59ca-4c55-aaf7-caab8965d2bb/proxy-httpd/0.log" Dec 10 15:31:26 crc kubenswrapper[4847]: I1210 15:31:26.930686 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hkv2r_daaeae82-74d4-4d18-8178-58ca7246b605/swift-ring-rebalance/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.078865 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-auditor/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.102140 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-reaper/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.254853 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-replicator/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.281535 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/account-server/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.314575 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-replicator/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.370043 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-auditor/0.log" Dec 10 15:31:27 crc kubenswrapper[4847]: I1210 15:31:27.610177 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-server/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.008656 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-expirer/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.021817 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-auditor/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.042528 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/container-updater/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.108857 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-replicator/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.170670 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-server/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.250743 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/rsync/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.305094 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/object-updater/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.347877 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_88c96cdf-8002-4829-9f8c-3abfc8315722/swift-recon-cron/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.514785 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-vn9kp_f8a7c3e5-a3f9-41aa-ab16-2e89d3a63b55/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.639907 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1240adcf-343f-4c3e-8323-d4ca80e2b99d/tempest-tests-tempest-tests-runner/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.785863 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a3d5c6e1-891d-4aa5-a9c5-ccaaaee1dad7/test-operator-logs-container/0.log" Dec 10 15:31:28 crc kubenswrapper[4847]: I1210 15:31:28.905292 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ckcqr_7a627a8d-ac7d-4d5d-9258-3342ca59930d/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 15:31:29 crc kubenswrapper[4847]: I1210 15:31:29.757699 4847 generic.go:334] "Generic (PLEG): container finished" podID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerID="e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09" exitCode=0 Dec 10 15:31:29 crc kubenswrapper[4847]: I1210 15:31:29.757773 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerDied","Data":"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09"} Dec 10 15:31:31 crc kubenswrapper[4847]: I1210 15:31:31.012167 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:31:31 crc kubenswrapper[4847]: I1210 15:31:31.012826 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:31:31 crc kubenswrapper[4847]: I1210 15:31:31.804032 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerStarted","Data":"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec"} Dec 10 15:31:31 crc kubenswrapper[4847]: I1210 15:31:31.850126 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qcncf" podStartSLOduration=3.861522344 podStartE2EDuration="10.850104263s" podCreationTimestamp="2025-12-10 15:31:21 +0000 UTC" firstStartedPulling="2025-12-10 15:31:23.671053766 +0000 UTC m=+4033.240271396" lastFinishedPulling="2025-12-10 15:31:30.659635685 +0000 UTC m=+4040.228853315" observedRunningTime="2025-12-10 15:31:31.832186349 +0000 UTC m=+4041.401403979" watchObservedRunningTime="2025-12-10 15:31:31.850104263 +0000 UTC m=+4041.419321893" Dec 10 15:31:32 crc kubenswrapper[4847]: I1210 15:31:32.162756 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:32 crc kubenswrapper[4847]: I1210 15:31:32.162800 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:33 crc kubenswrapper[4847]: I1210 15:31:33.215408 4847 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qcncf" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="registry-server" probeResult="failure" output=< Dec 10 15:31:33 crc kubenswrapper[4847]: timeout: failed to connect service ":50051" within 1s Dec 10 15:31:33 crc kubenswrapper[4847]: > Dec 10 15:31:37 crc kubenswrapper[4847]: I1210 15:31:37.096869 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fc672732-248d-4f66-88fd-3b98ce15e78d/memcached/0.log" Dec 10 15:31:42 crc kubenswrapper[4847]: I1210 15:31:42.214787 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:42 crc kubenswrapper[4847]: I1210 15:31:42.274932 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:42 crc kubenswrapper[4847]: I1210 15:31:42.451433 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:43 crc kubenswrapper[4847]: I1210 15:31:43.903465 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qcncf" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="registry-server" containerID="cri-o://bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec" gracePeriod=2 Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.393096 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.468917 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6bkg\" (UniqueName: \"kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg\") pod \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.469325 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content\") pod \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.469436 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities\") pod \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\" (UID: \"fea5660d-5b3b-42b4-9d36-8ba688ad2075\") " Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.470303 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities" (OuterVolumeSpecName: "utilities") pod "fea5660d-5b3b-42b4-9d36-8ba688ad2075" (UID: "fea5660d-5b3b-42b4-9d36-8ba688ad2075"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.480316 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg" (OuterVolumeSpecName: "kube-api-access-p6bkg") pod "fea5660d-5b3b-42b4-9d36-8ba688ad2075" (UID: "fea5660d-5b3b-42b4-9d36-8ba688ad2075"). InnerVolumeSpecName "kube-api-access-p6bkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.571568 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6bkg\" (UniqueName: \"kubernetes.io/projected/fea5660d-5b3b-42b4-9d36-8ba688ad2075-kube-api-access-p6bkg\") on node \"crc\" DevicePath \"\"" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.571850 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.588283 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fea5660d-5b3b-42b4-9d36-8ba688ad2075" (UID: "fea5660d-5b3b-42b4-9d36-8ba688ad2075"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.673596 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fea5660d-5b3b-42b4-9d36-8ba688ad2075-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.912873 4847 generic.go:334] "Generic (PLEG): container finished" podID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerID="bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec" exitCode=0 Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.912915 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerDied","Data":"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec"} Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.912944 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcncf" event={"ID":"fea5660d-5b3b-42b4-9d36-8ba688ad2075","Type":"ContainerDied","Data":"bee3d4ce7c606106898f50915900a41a47a035a650210f08d556195a30b58964"} Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.912965 4847 scope.go:117] "RemoveContainer" containerID="bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.913905 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcncf" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.941078 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.941273 4847 scope.go:117] "RemoveContainer" containerID="e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09" Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.948918 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qcncf"] Dec 10 15:31:44 crc kubenswrapper[4847]: I1210 15:31:44.965310 4847 scope.go:117] "RemoveContainer" containerID="038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.024452 4847 scope.go:117] "RemoveContainer" containerID="bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec" Dec 10 15:31:45 crc kubenswrapper[4847]: E1210 15:31:45.025002 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec\": container with ID starting with bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec not found: ID does not exist" containerID="bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.025057 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec"} err="failed to get container status \"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec\": rpc error: code = NotFound desc = could not find container \"bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec\": container with ID starting with bdd14d585d2d3a66f61e2f767a6a537e416156a413be2a6c7640be1b5f815cec not found: ID does not exist" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.025086 4847 scope.go:117] "RemoveContainer" containerID="e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09" Dec 10 15:31:45 crc kubenswrapper[4847]: E1210 15:31:45.025476 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09\": container with ID starting with e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09 not found: ID does not exist" containerID="e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.025588 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09"} err="failed to get container status \"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09\": rpc error: code = NotFound desc = could not find container \"e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09\": container with ID starting with e9fcba896743c81a649a3340a9fd52aa506cfec46f158378d15bdc1c61a44f09 not found: ID does not exist" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.025662 4847 scope.go:117] "RemoveContainer" containerID="038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b" Dec 10 15:31:45 crc kubenswrapper[4847]: E1210 15:31:45.025997 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b\": container with ID starting with 038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b not found: ID does not exist" containerID="038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b" Dec 10 15:31:45 crc kubenswrapper[4847]: I1210 15:31:45.026027 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b"} err="failed to get container status \"038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b\": rpc error: code = NotFound desc = could not find container \"038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b\": container with ID starting with 038f7e500db943a5ee7643fa3c7b5787b03671fa902cda2dc0c598d8f6aa391b not found: ID does not exist" Dec 10 15:31:46 crc kubenswrapper[4847]: I1210 15:31:46.773413 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" path="/var/lib/kubelet/pods/fea5660d-5b3b-42b4-9d36-8ba688ad2075/volumes" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.325701 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.537629 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.550675 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.557581 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.721998 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/pull/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.755400 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/extract/0.log" Dec 10 15:31:56 crc kubenswrapper[4847]: I1210 15:31:56.766098 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f900f5a499809861b79b53e47e6bd4bff2fbb933e2488cc54caeeefaaqzxgq_f5631548-ea6d-4a31-a0ae-622f343829fa/util/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.302293 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqjxj_80cfb202-f181-4c1a-b9d6-a7cd036b93db/kube-rbac-proxy/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.345789 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-rqjxj_80cfb202-f181-4c1a-b9d6-a7cd036b93db/manager/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.363111 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-pdp9r_df30724c-6ddb-4e68-8f57-31f8bb262678/kube-rbac-proxy/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.540291 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-pdp9r_df30724c-6ddb-4e68-8f57-31f8bb262678/manager/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.567519 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-pmnjz_3c7abfb3-9360-42df-86b1-127066817fe8/kube-rbac-proxy/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.570259 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-pmnjz_3c7abfb3-9360-42df-86b1-127066817fe8/manager/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.763590 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-fl8dl_519d2b7e-57ab-4f4e-a199-68b3fe9a16bf/kube-rbac-proxy/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.806750 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-fl8dl_519d2b7e-57ab-4f4e-a199-68b3fe9a16bf/manager/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.923469 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-t2wb8_defc8bda-15a6-4432-8079-9606a8fdb2e9/kube-rbac-proxy/0.log" Dec 10 15:31:57 crc kubenswrapper[4847]: I1210 15:31:57.951210 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-t2wb8_defc8bda-15a6-4432-8079-9606a8fdb2e9/manager/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.071549 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gdhjq_82fa6e27-3c61-4a46-a51c-1f9e034a870c/kube-rbac-proxy/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.169500 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gdhjq_82fa6e27-3c61-4a46-a51c-1f9e034a870c/manager/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.300406 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-twm4t_73f1dc78-cf06-4c8d-a57a-314ce95cc1d5/kube-rbac-proxy/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.368628 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4j2fk_3a197914-3847-44b5-aa44-527bef265cc5/kube-rbac-proxy/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.480394 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-twm4t_73f1dc78-cf06-4c8d-a57a-314ce95cc1d5/manager/0.log" Dec 10 15:31:58 crc kubenswrapper[4847]: I1210 15:31:58.556004 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4j2fk_3a197914-3847-44b5-aa44-527bef265cc5/manager/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.205489 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-55zkq_2fa5c5a1-e3d6-424c-b52a-d01af15458cf/kube-rbac-proxy/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.217291 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-q8fkp_161aaad4-8b14-4a9c-b793-a7937e078aef/kube-rbac-proxy/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.248514 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-55zkq_2fa5c5a1-e3d6-424c-b52a-d01af15458cf/manager/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.402518 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-sptmg_9f72b089-2518-4144-a558-bfb5c490d594/kube-rbac-proxy/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.409474 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-q8fkp_161aaad4-8b14-4a9c-b793-a7937e078aef/manager/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.442420 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-sptmg_9f72b089-2518-4144-a558-bfb5c490d594/manager/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.732493 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gjtlq_3654ef6c-9eef-40ec-acd3-c101a817f345/kube-rbac-proxy/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.822944 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gjtlq_3654ef6c-9eef-40ec-acd3-c101a817f345/manager/0.log" Dec 10 15:31:59 crc kubenswrapper[4847]: I1210 15:31:59.929233 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-kds4v_79090a78-3da6-4b17-a875-fc8c2d534076/kube-rbac-proxy/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.011977 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c96c4_db714bfd-5854-4a09-a6d0-acf7e3c539a4/kube-rbac-proxy/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.062506 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-kds4v_79090a78-3da6-4b17-a875-fc8c2d534076/manager/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.077989 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c96c4_db714bfd-5854-4a09-a6d0-acf7e3c539a4/manager/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.200773 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f2swnv_d78cab92-e99d-41fb-a35d-ff63d3ceba23/kube-rbac-proxy/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.244381 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f2swnv_d78cab92-e99d-41fb-a35d-ff63d3ceba23/manager/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.612118 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-ptjmg_692f8962-1682-4a0d-b693-24e90e81bb17/registry-server/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.742580 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6646c79bbd-rqsnh_caaff4fe-5b58-4a77-b3fe-217742051274/operator/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.833229 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sbtv2_64a44edf-788e-4286-83af-6072b8d8b437/manager/0.log" Dec 10 15:32:00 crc kubenswrapper[4847]: I1210 15:32:00.857021 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sbtv2_64a44edf-788e-4286-83af-6072b8d8b437/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.014419 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.014799 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.027179 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-t224h_87dac2ce-ec87-4d12-9ac8-0a79bedbfa62/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.071601 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-t224h_87dac2ce-ec87-4d12-9ac8-0a79bedbfa62/manager/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.168112 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-d5mdv_4706d44c-19de-49a6-b604-dbe911592e48/operator/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.345304 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-4dfs5_9b0a2f2c-971b-40cc-8544-9842f7878d62/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.407986 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-4dfs5_9b0a2f2c-971b-40cc-8544-9842f7878d62/manager/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.443847 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79b6c4fddb-8mzvt_b1d011f5-5052-431d-9933-314f65930ce4/manager/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.539392 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-z6f9c_971c05b8-38ac-411a-bdc8-b40ed70daab5/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.569392 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-z6f9c_971c05b8-38ac-411a-bdc8-b40ed70daab5/manager/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.653226 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8k92s_c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.657466 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8k92s_c1bee1d7-7050-4ef2-84f3-5fe2fbe3fc90/manager/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.778517 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-4dhdf_0c5922f3-c544-4993-b978-4a035ec08e1b/kube-rbac-proxy/0.log" Dec 10 15:32:01 crc kubenswrapper[4847]: I1210 15:32:01.818687 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-4dhdf_0c5922f3-c544-4993-b978-4a035ec08e1b/manager/0.log" Dec 10 15:32:20 crc kubenswrapper[4847]: I1210 15:32:20.242879 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-wx4gj_89217e6c-9f20-45f0-8a26-7c808efc3359/control-plane-machine-set-operator/0.log" Dec 10 15:32:20 crc kubenswrapper[4847]: I1210 15:32:20.460171 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6lxgl_57fb36d2-42fe-4993-a454-1db53581dc35/kube-rbac-proxy/0.log" Dec 10 15:32:20 crc kubenswrapper[4847]: I1210 15:32:20.460888 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6lxgl_57fb36d2-42fe-4993-a454-1db53581dc35/machine-api-operator/0.log" Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.010782 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.011415 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.011480 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.012513 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.012608 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858" gracePeriod=600 Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.333536 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858" exitCode=0 Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.333573 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858"} Dec 10 15:32:31 crc kubenswrapper[4847]: I1210 15:32:31.333993 4847 scope.go:117] "RemoveContainer" containerID="51fc01a3c1cd18604b988c66d29b69a2927d0cb8ce027e5e77a6540a72c0e65c" Dec 10 15:32:32 crc kubenswrapper[4847]: I1210 15:32:32.343742 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerStarted","Data":"ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b"} Dec 10 15:32:33 crc kubenswrapper[4847]: I1210 15:32:33.293449 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6nfcw_f0cacf0f-d6c7-45c1-9241-7a3eb577594c/cert-manager-controller/0.log" Dec 10 15:32:33 crc kubenswrapper[4847]: I1210 15:32:33.506077 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-g8v52_ba5f50e7-82cb-435d-a7eb-5a0e3a181d0d/cert-manager-cainjector/0.log" Dec 10 15:32:33 crc kubenswrapper[4847]: I1210 15:32:33.518331 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7h9v6_2266f4a5-b282-4686-9a2c-e51d538a11cd/cert-manager-webhook/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.178283 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-tjd2c_d5575bcb-0d29-4538-8213-99a007548880/nmstate-console-plugin/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.211666 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wxpbs_b51a4da1-a921-414f-9985-044264a489ad/nmstate-handler/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.353204 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-5rj2v_e7bb30c3-5f7b-465a-8852-388f8204277b/nmstate-metrics/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.357216 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-5rj2v_e7bb30c3-5f7b-465a-8852-388f8204277b/kube-rbac-proxy/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.526100 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-x4fhc_64363456-a381-49ba-98f2-b16b9928b65b/nmstate-operator/0.log" Dec 10 15:32:46 crc kubenswrapper[4847]: I1210 15:32:46.559623 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4kbcp_ef0ffd00-af07-4843-9b34-2cf613252d7c/nmstate-webhook/0.log" Dec 10 15:33:01 crc kubenswrapper[4847]: I1210 15:33:01.658443 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6mjl9_d00cc0ff-5b61-4786-abd8-44da90bdb484/kube-rbac-proxy/0.log" Dec 10 15:33:01 crc kubenswrapper[4847]: I1210 15:33:01.795469 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6mjl9_d00cc0ff-5b61-4786-abd8-44da90bdb484/controller/0.log" Dec 10 15:33:01 crc kubenswrapper[4847]: I1210 15:33:01.829288 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.036175 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.068584 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.079816 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.116792 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.259562 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.266362 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.272109 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.319028 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.492035 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-reloader/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.506574 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-frr-files/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.520618 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/cp-metrics/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.523540 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/controller/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.660833 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/frr-metrics/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.715667 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/kube-rbac-proxy/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.756287 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/kube-rbac-proxy-frr/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.878074 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/reloader/0.log" Dec 10 15:33:02 crc kubenswrapper[4847]: I1210 15:33:02.994824 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-ng6j2_ec15bd80-1069-4aed-9f51-43140a7f6c5e/frr-k8s-webhook-server/0.log" Dec 10 15:33:03 crc kubenswrapper[4847]: I1210 15:33:03.175970 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b5f976d54-lrrhd_12bcc429-a2dd-4b92-a99b-86ea695e6147/manager/0.log" Dec 10 15:33:03 crc kubenswrapper[4847]: I1210 15:33:03.326543 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7bc66b7c4c-gxscv_528aeb2a-2b6f-4e7b-8a22-b40ecca39c45/webhook-server/0.log" Dec 10 15:33:04 crc kubenswrapper[4847]: I1210 15:33:04.081376 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-96l7c_b837e40a-6332-4206-94f5-2555dc1cd40c/frr/0.log" Dec 10 15:33:04 crc kubenswrapper[4847]: I1210 15:33:04.271826 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7bf6v_74b7d233-5faf-4997-9986-f6fb167df94b/kube-rbac-proxy/0.log" Dec 10 15:33:04 crc kubenswrapper[4847]: I1210 15:33:04.551454 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7bf6v_74b7d233-5faf-4997-9986-f6fb167df94b/speaker/0.log" Dec 10 15:33:17 crc kubenswrapper[4847]: I1210 15:33:17.870477 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:33:17 crc kubenswrapper[4847]: I1210 15:33:17.994194 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.021548 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.112924 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.197325 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/util/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.223928 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/extract/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.225442 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f4n2hz_b5094984-5419-4b07-bbfd-033eb535c588/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.360312 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.505708 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.520878 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.524272 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.702863 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/extract/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.711897 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/pull/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.716804 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83mqwrm_113a6c40-884a-41c9-9019-6ad88c99e989/util/0.log" Dec 10 15:33:18 crc kubenswrapper[4847]: I1210 15:33:18.869282 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.025205 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.081569 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.105244 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.228278 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.265675 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/extract-utilities/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.444588 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.528471 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kqks9_2f6069b4-37e0-452f-9361-f5da5a9a0f52/registry-server/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.619256 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.630960 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.664027 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.824971 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-content/0.log" Dec 10 15:33:19 crc kubenswrapper[4847]: I1210 15:33:19.854890 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/extract-utilities/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.106656 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dz9r4_f55386e3-acde-49bd-941e-a92591ec95de/marketplace-operator/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.114086 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.359863 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.368227 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7zpg_ad6ec978-5852-4454-a0ef-b4d75cbe12d2/registry-server/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.372993 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.399991 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.557963 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-content/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.585184 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/extract-utilities/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.795265 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.804038 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4frbk_ea23c9a6-79c4-48c1-bb39-2c75dc805ea8/registry-server/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.973902 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:33:20 crc kubenswrapper[4847]: I1210 15:33:20.989330 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:33:21 crc kubenswrapper[4847]: I1210 15:33:21.005449 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:33:21 crc kubenswrapper[4847]: I1210 15:33:21.200691 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-utilities/0.log" Dec 10 15:33:21 crc kubenswrapper[4847]: I1210 15:33:21.206335 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/extract-content/0.log" Dec 10 15:33:21 crc kubenswrapper[4847]: I1210 15:33:21.673543 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-djwns_902cf9a3-e8c7-46c9-b058-de3249e6ce65/registry-server/0.log" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.465796 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:43 crc kubenswrapper[4847]: E1210 15:33:43.466731 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="extract-utilities" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.466744 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="extract-utilities" Dec 10 15:33:43 crc kubenswrapper[4847]: E1210 15:33:43.466759 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="registry-server" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.466765 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="registry-server" Dec 10 15:33:43 crc kubenswrapper[4847]: E1210 15:33:43.466786 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="extract-content" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.466792 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="extract-content" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.466976 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="fea5660d-5b3b-42b4-9d36-8ba688ad2075" containerName="registry-server" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.485734 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.497262 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.589909 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-787n2\" (UniqueName: \"kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.589978 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.590020 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.691977 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-787n2\" (UniqueName: \"kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.692058 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.692101 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.692588 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.692699 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.723505 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-787n2\" (UniqueName: \"kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2\") pod \"community-operators-vnn7k\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:43 crc kubenswrapper[4847]: I1210 15:33:43.833324 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:44 crc kubenswrapper[4847]: I1210 15:33:44.272383 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:44 crc kubenswrapper[4847]: I1210 15:33:44.982746 4847 generic.go:334] "Generic (PLEG): container finished" podID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerID="02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224" exitCode=0 Dec 10 15:33:44 crc kubenswrapper[4847]: I1210 15:33:44.982783 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerDied","Data":"02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224"} Dec 10 15:33:44 crc kubenswrapper[4847]: I1210 15:33:44.983137 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerStarted","Data":"211b4ae3cbe4ae02f963ca962bba014679358ae7aefd924aee1bdcf4e93833af"} Dec 10 15:33:44 crc kubenswrapper[4847]: I1210 15:33:44.984631 4847 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 15:33:47 crc kubenswrapper[4847]: I1210 15:33:47.007882 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerStarted","Data":"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7"} Dec 10 15:33:48 crc kubenswrapper[4847]: I1210 15:33:48.020217 4847 generic.go:334] "Generic (PLEG): container finished" podID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerID="813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7" exitCode=0 Dec 10 15:33:48 crc kubenswrapper[4847]: I1210 15:33:48.020290 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerDied","Data":"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7"} Dec 10 15:33:50 crc kubenswrapper[4847]: I1210 15:33:50.039053 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerStarted","Data":"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c"} Dec 10 15:33:50 crc kubenswrapper[4847]: I1210 15:33:50.065409 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vnn7k" podStartSLOduration=2.645557971 podStartE2EDuration="7.065390986s" podCreationTimestamp="2025-12-10 15:33:43 +0000 UTC" firstStartedPulling="2025-12-10 15:33:44.984337291 +0000 UTC m=+4174.553554921" lastFinishedPulling="2025-12-10 15:33:49.404170156 +0000 UTC m=+4178.973387936" observedRunningTime="2025-12-10 15:33:50.057576626 +0000 UTC m=+4179.626794256" watchObservedRunningTime="2025-12-10 15:33:50.065390986 +0000 UTC m=+4179.634608616" Dec 10 15:33:53 crc kubenswrapper[4847]: I1210 15:33:53.834312 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:53 crc kubenswrapper[4847]: I1210 15:33:53.835349 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:53 crc kubenswrapper[4847]: I1210 15:33:53.912478 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:54 crc kubenswrapper[4847]: I1210 15:33:54.127006 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:54 crc kubenswrapper[4847]: I1210 15:33:54.175865 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.093075 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vnn7k" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="registry-server" containerID="cri-o://1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c" gracePeriod=2 Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.852266 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.946304 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities\") pod \"479c210a-ffd1-4c55-80cc-fea313eccd7a\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.947453 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities" (OuterVolumeSpecName: "utilities") pod "479c210a-ffd1-4c55-80cc-fea313eccd7a" (UID: "479c210a-ffd1-4c55-80cc-fea313eccd7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.948371 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content\") pod \"479c210a-ffd1-4c55-80cc-fea313eccd7a\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.948498 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-787n2\" (UniqueName: \"kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2\") pod \"479c210a-ffd1-4c55-80cc-fea313eccd7a\" (UID: \"479c210a-ffd1-4c55-80cc-fea313eccd7a\") " Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.950001 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:33:56 crc kubenswrapper[4847]: I1210 15:33:56.955757 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2" (OuterVolumeSpecName: "kube-api-access-787n2") pod "479c210a-ffd1-4c55-80cc-fea313eccd7a" (UID: "479c210a-ffd1-4c55-80cc-fea313eccd7a"). InnerVolumeSpecName "kube-api-access-787n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.004957 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "479c210a-ffd1-4c55-80cc-fea313eccd7a" (UID: "479c210a-ffd1-4c55-80cc-fea313eccd7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.051797 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/479c210a-ffd1-4c55-80cc-fea313eccd7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.051843 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-787n2\" (UniqueName: \"kubernetes.io/projected/479c210a-ffd1-4c55-80cc-fea313eccd7a-kube-api-access-787n2\") on node \"crc\" DevicePath \"\"" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.102104 4847 generic.go:334] "Generic (PLEG): container finished" podID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerID="1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c" exitCode=0 Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.102150 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerDied","Data":"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c"} Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.102178 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vnn7k" event={"ID":"479c210a-ffd1-4c55-80cc-fea313eccd7a","Type":"ContainerDied","Data":"211b4ae3cbe4ae02f963ca962bba014679358ae7aefd924aee1bdcf4e93833af"} Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.102198 4847 scope.go:117] "RemoveContainer" containerID="1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.102320 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vnn7k" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.138212 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.141909 4847 scope.go:117] "RemoveContainer" containerID="813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.147187 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vnn7k"] Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.161100 4847 scope.go:117] "RemoveContainer" containerID="02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.206601 4847 scope.go:117] "RemoveContainer" containerID="1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c" Dec 10 15:33:57 crc kubenswrapper[4847]: E1210 15:33:57.207021 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c\": container with ID starting with 1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c not found: ID does not exist" containerID="1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.207052 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c"} err="failed to get container status \"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c\": rpc error: code = NotFound desc = could not find container \"1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c\": container with ID starting with 1826f6aabc4e158deffda32a4c060ec6cc05e6544cb0cb1ffdae51238a41a89c not found: ID does not exist" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.207072 4847 scope.go:117] "RemoveContainer" containerID="813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7" Dec 10 15:33:57 crc kubenswrapper[4847]: E1210 15:33:57.207291 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7\": container with ID starting with 813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7 not found: ID does not exist" containerID="813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.207312 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7"} err="failed to get container status \"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7\": rpc error: code = NotFound desc = could not find container \"813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7\": container with ID starting with 813c0c395869026394050079fbaa01c25f4ba2b640fd944ee3aac2d4e7f19fc7 not found: ID does not exist" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.207324 4847 scope.go:117] "RemoveContainer" containerID="02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224" Dec 10 15:33:57 crc kubenswrapper[4847]: E1210 15:33:57.207625 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224\": container with ID starting with 02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224 not found: ID does not exist" containerID="02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224" Dec 10 15:33:57 crc kubenswrapper[4847]: I1210 15:33:57.207675 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224"} err="failed to get container status \"02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224\": rpc error: code = NotFound desc = could not find container \"02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224\": container with ID starting with 02bcf94860b80a0c3efb12b801d2e27e7be1ac59a6737da4af8910d538165224 not found: ID does not exist" Dec 10 15:33:58 crc kubenswrapper[4847]: I1210 15:33:58.774391 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" path="/var/lib/kubelet/pods/479c210a-ffd1-4c55-80cc-fea313eccd7a/volumes" Dec 10 15:34:31 crc kubenswrapper[4847]: I1210 15:34:31.012613 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:34:31 crc kubenswrapper[4847]: I1210 15:34:31.013385 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:35:01 crc kubenswrapper[4847]: I1210 15:35:01.011723 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:35:01 crc kubenswrapper[4847]: I1210 15:35:01.012481 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:35:03 crc kubenswrapper[4847]: I1210 15:35:03.784630 4847 generic.go:334] "Generic (PLEG): container finished" podID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerID="f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3" exitCode=0 Dec 10 15:35:03 crc kubenswrapper[4847]: I1210 15:35:03.784737 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5jsgq/must-gather-q76lt" event={"ID":"5233e69a-8779-4a5a-945f-3acf0bcf4334","Type":"ContainerDied","Data":"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3"} Dec 10 15:35:03 crc kubenswrapper[4847]: I1210 15:35:03.786341 4847 scope.go:117] "RemoveContainer" containerID="f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3" Dec 10 15:35:04 crc kubenswrapper[4847]: I1210 15:35:04.334428 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jsgq_must-gather-q76lt_5233e69a-8779-4a5a-945f-3acf0bcf4334/gather/0.log" Dec 10 15:35:14 crc kubenswrapper[4847]: I1210 15:35:14.801406 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5jsgq/must-gather-q76lt"] Dec 10 15:35:14 crc kubenswrapper[4847]: I1210 15:35:14.803243 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5jsgq/must-gather-q76lt" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="copy" containerID="cri-o://ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6" gracePeriod=2 Dec 10 15:35:14 crc kubenswrapper[4847]: I1210 15:35:14.811856 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5jsgq/must-gather-q76lt"] Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.242682 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jsgq_must-gather-q76lt_5233e69a-8779-4a5a-945f-3acf0bcf4334/copy/0.log" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.244205 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.405792 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output\") pod \"5233e69a-8779-4a5a-945f-3acf0bcf4334\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.406054 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c76kt\" (UniqueName: \"kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt\") pod \"5233e69a-8779-4a5a-945f-3acf0bcf4334\" (UID: \"5233e69a-8779-4a5a-945f-3acf0bcf4334\") " Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.418936 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt" (OuterVolumeSpecName: "kube-api-access-c76kt") pod "5233e69a-8779-4a5a-945f-3acf0bcf4334" (UID: "5233e69a-8779-4a5a-945f-3acf0bcf4334"). InnerVolumeSpecName "kube-api-access-c76kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.508910 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c76kt\" (UniqueName: \"kubernetes.io/projected/5233e69a-8779-4a5a-945f-3acf0bcf4334-kube-api-access-c76kt\") on node \"crc\" DevicePath \"\"" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.565040 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5233e69a-8779-4a5a-945f-3acf0bcf4334" (UID: "5233e69a-8779-4a5a-945f-3acf0bcf4334"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.611657 4847 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5233e69a-8779-4a5a-945f-3acf0bcf4334-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.903797 4847 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5jsgq_must-gather-q76lt_5233e69a-8779-4a5a-945f-3acf0bcf4334/copy/0.log" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.904293 4847 generic.go:334] "Generic (PLEG): container finished" podID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerID="ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6" exitCode=143 Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.904338 4847 scope.go:117] "RemoveContainer" containerID="ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.904393 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5jsgq/must-gather-q76lt" Dec 10 15:35:15 crc kubenswrapper[4847]: I1210 15:35:15.926600 4847 scope.go:117] "RemoveContainer" containerID="f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3" Dec 10 15:35:16 crc kubenswrapper[4847]: I1210 15:35:16.005656 4847 scope.go:117] "RemoveContainer" containerID="ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6" Dec 10 15:35:16 crc kubenswrapper[4847]: E1210 15:35:16.006023 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6\": container with ID starting with ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6 not found: ID does not exist" containerID="ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6" Dec 10 15:35:16 crc kubenswrapper[4847]: I1210 15:35:16.006067 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6"} err="failed to get container status \"ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6\": rpc error: code = NotFound desc = could not find container \"ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6\": container with ID starting with ede1e8b56417be2975f57bad77467a2faa5db9a79964ffe81ee1dd5e6fa0e1c6 not found: ID does not exist" Dec 10 15:35:16 crc kubenswrapper[4847]: I1210 15:35:16.006094 4847 scope.go:117] "RemoveContainer" containerID="f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3" Dec 10 15:35:16 crc kubenswrapper[4847]: E1210 15:35:16.006381 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3\": container with ID starting with f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3 not found: ID does not exist" containerID="f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3" Dec 10 15:35:16 crc kubenswrapper[4847]: I1210 15:35:16.006428 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3"} err="failed to get container status \"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3\": rpc error: code = NotFound desc = could not find container \"f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3\": container with ID starting with f926c7dc22772f29fef9b1a9dfe1359ecc4a316f8bcbe1e41e0494e6c579aff3 not found: ID does not exist" Dec 10 15:35:16 crc kubenswrapper[4847]: I1210 15:35:16.772657 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" path="/var/lib/kubelet/pods/5233e69a-8779-4a5a-945f-3acf0bcf4334/volumes" Dec 10 15:35:31 crc kubenswrapper[4847]: I1210 15:35:31.011560 4847 patch_prober.go:28] interesting pod/machine-config-daemon-gvdwq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 15:35:31 crc kubenswrapper[4847]: I1210 15:35:31.012499 4847 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 15:35:31 crc kubenswrapper[4847]: I1210 15:35:31.012586 4847 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" Dec 10 15:35:31 crc kubenswrapper[4847]: I1210 15:35:31.014070 4847 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b"} pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 15:35:31 crc kubenswrapper[4847]: I1210 15:35:31.014212 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerName="machine-config-daemon" containerID="cri-o://ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" gracePeriod=600 Dec 10 15:35:31 crc kubenswrapper[4847]: E1210 15:35:31.147530 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:35:32 crc kubenswrapper[4847]: I1210 15:35:32.077384 4847 generic.go:334] "Generic (PLEG): container finished" podID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" exitCode=0 Dec 10 15:35:32 crc kubenswrapper[4847]: I1210 15:35:32.077455 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" event={"ID":"8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324","Type":"ContainerDied","Data":"ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b"} Dec 10 15:35:32 crc kubenswrapper[4847]: I1210 15:35:32.077799 4847 scope.go:117] "RemoveContainer" containerID="c5bd651efe3c2cd8e092a074c0f6e049f388d32b0e1d7da9d63296b78f2a4858" Dec 10 15:35:32 crc kubenswrapper[4847]: I1210 15:35:32.078723 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:35:32 crc kubenswrapper[4847]: E1210 15:35:32.079017 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:35:44 crc kubenswrapper[4847]: I1210 15:35:44.760220 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:35:44 crc kubenswrapper[4847]: E1210 15:35:44.761026 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:35:58 crc kubenswrapper[4847]: I1210 15:35:58.759575 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:35:58 crc kubenswrapper[4847]: E1210 15:35:58.760260 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:36:11 crc kubenswrapper[4847]: I1210 15:36:11.760352 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:36:11 crc kubenswrapper[4847]: E1210 15:36:11.762107 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:36:25 crc kubenswrapper[4847]: I1210 15:36:25.705341 4847 scope.go:117] "RemoveContainer" containerID="35f176381fdd531eb5d71f21a07b5264f780bfb4bd53a7bca353ec0890ce8f06" Dec 10 15:36:25 crc kubenswrapper[4847]: I1210 15:36:25.759526 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:36:25 crc kubenswrapper[4847]: E1210 15:36:25.759875 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:36:37 crc kubenswrapper[4847]: I1210 15:36:37.759464 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:36:37 crc kubenswrapper[4847]: E1210 15:36:37.760258 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:36:50 crc kubenswrapper[4847]: I1210 15:36:50.759926 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:36:50 crc kubenswrapper[4847]: E1210 15:36:50.761026 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:37:01 crc kubenswrapper[4847]: I1210 15:37:01.759883 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:37:01 crc kubenswrapper[4847]: E1210 15:37:01.760914 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:37:15 crc kubenswrapper[4847]: I1210 15:37:15.760469 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:37:15 crc kubenswrapper[4847]: E1210 15:37:15.761367 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:37:30 crc kubenswrapper[4847]: I1210 15:37:30.766814 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:37:30 crc kubenswrapper[4847]: E1210 15:37:30.767544 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:37:41 crc kubenswrapper[4847]: I1210 15:37:41.759856 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:37:41 crc kubenswrapper[4847]: E1210 15:37:41.760741 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:37:52 crc kubenswrapper[4847]: I1210 15:37:52.760263 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:37:52 crc kubenswrapper[4847]: E1210 15:37:52.761235 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:38:06 crc kubenswrapper[4847]: I1210 15:38:06.765326 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:38:06 crc kubenswrapper[4847]: E1210 15:38:06.766360 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.184107 4847 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:14 crc kubenswrapper[4847]: E1210 15:38:14.185302 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="copy" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185317 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="copy" Dec 10 15:38:14 crc kubenswrapper[4847]: E1210 15:38:14.185335 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="extract-utilities" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185343 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="extract-utilities" Dec 10 15:38:14 crc kubenswrapper[4847]: E1210 15:38:14.185372 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="extract-content" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185380 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="extract-content" Dec 10 15:38:14 crc kubenswrapper[4847]: E1210 15:38:14.185396 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="gather" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185404 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="gather" Dec 10 15:38:14 crc kubenswrapper[4847]: E1210 15:38:14.185426 4847 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="registry-server" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185433 4847 state_mem.go:107] "Deleted CPUSet assignment" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="registry-server" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185641 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="copy" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185660 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="5233e69a-8779-4a5a-945f-3acf0bcf4334" containerName="gather" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.185684 4847 memory_manager.go:354] "RemoveStaleState removing state" podUID="479c210a-ffd1-4c55-80cc-fea313eccd7a" containerName="registry-server" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.187502 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.212802 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.323160 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptgdk\" (UniqueName: \"kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.323369 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.323654 4847 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.425812 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptgdk\" (UniqueName: \"kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.425945 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.426031 4847 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.426484 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.426500 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.463162 4847 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptgdk\" (UniqueName: \"kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk\") pod \"certified-operators-hsrtq\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:14 crc kubenswrapper[4847]: I1210 15:38:14.519935 4847 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:15 crc kubenswrapper[4847]: I1210 15:38:15.069287 4847 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:15 crc kubenswrapper[4847]: W1210 15:38:15.104782 4847 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f9f5bc0_dba0_4c11_a612_5ab327ed22b2.slice/crio-42ce366b574660e75e678ae4dc693e0a27e218d7bf16b0170a7d1fdded6362ba WatchSource:0}: Error finding container 42ce366b574660e75e678ae4dc693e0a27e218d7bf16b0170a7d1fdded6362ba: Status 404 returned error can't find the container with id 42ce366b574660e75e678ae4dc693e0a27e218d7bf16b0170a7d1fdded6362ba Dec 10 15:38:15 crc kubenswrapper[4847]: I1210 15:38:15.658296 4847 generic.go:334] "Generic (PLEG): container finished" podID="2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" containerID="d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b" exitCode=0 Dec 10 15:38:15 crc kubenswrapper[4847]: I1210 15:38:15.658782 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerDied","Data":"d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b"} Dec 10 15:38:15 crc kubenswrapper[4847]: I1210 15:38:15.659662 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerStarted","Data":"42ce366b574660e75e678ae4dc693e0a27e218d7bf16b0170a7d1fdded6362ba"} Dec 10 15:38:16 crc kubenswrapper[4847]: I1210 15:38:16.674811 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerStarted","Data":"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f"} Dec 10 15:38:17 crc kubenswrapper[4847]: I1210 15:38:17.688266 4847 generic.go:334] "Generic (PLEG): container finished" podID="2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" containerID="1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f" exitCode=0 Dec 10 15:38:17 crc kubenswrapper[4847]: I1210 15:38:17.688354 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerDied","Data":"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f"} Dec 10 15:38:20 crc kubenswrapper[4847]: I1210 15:38:20.715856 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerStarted","Data":"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9"} Dec 10 15:38:20 crc kubenswrapper[4847]: I1210 15:38:20.749312 4847 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hsrtq" podStartSLOduration=2.34182128 podStartE2EDuration="6.74928539s" podCreationTimestamp="2025-12-10 15:38:14 +0000 UTC" firstStartedPulling="2025-12-10 15:38:15.662544813 +0000 UTC m=+4445.231762443" lastFinishedPulling="2025-12-10 15:38:20.070008923 +0000 UTC m=+4449.639226553" observedRunningTime="2025-12-10 15:38:20.740610698 +0000 UTC m=+4450.309828328" watchObservedRunningTime="2025-12-10 15:38:20.74928539 +0000 UTC m=+4450.318503020" Dec 10 15:38:21 crc kubenswrapper[4847]: I1210 15:38:21.760604 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:38:21 crc kubenswrapper[4847]: E1210 15:38:21.761056 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324" Dec 10 15:38:24 crc kubenswrapper[4847]: I1210 15:38:24.520796 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:24 crc kubenswrapper[4847]: I1210 15:38:24.521199 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:25 crc kubenswrapper[4847]: I1210 15:38:25.262492 4847 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:25 crc kubenswrapper[4847]: I1210 15:38:25.326665 4847 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:25 crc kubenswrapper[4847]: I1210 15:38:25.506470 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:26 crc kubenswrapper[4847]: I1210 15:38:26.775020 4847 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hsrtq" podUID="2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" containerName="registry-server" containerID="cri-o://c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9" gracePeriod=2 Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.540947 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.610193 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content\") pod \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.610397 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities\") pod \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.610427 4847 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptgdk\" (UniqueName: \"kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk\") pod \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\" (UID: \"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2\") " Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.611623 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities" (OuterVolumeSpecName: "utilities") pod "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" (UID: "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.618062 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk" (OuterVolumeSpecName: "kube-api-access-ptgdk") pod "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" (UID: "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2"). InnerVolumeSpecName "kube-api-access-ptgdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.664666 4847 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" (UID: "2f9f5bc0-dba0-4c11-a612-5ab327ed22b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.712528 4847 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.712566 4847 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.712576 4847 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptgdk\" (UniqueName: \"kubernetes.io/projected/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2-kube-api-access-ptgdk\") on node \"crc\" DevicePath \"\"" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.793840 4847 generic.go:334] "Generic (PLEG): container finished" podID="2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" containerID="c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9" exitCode=0 Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.793885 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerDied","Data":"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9"} Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.793918 4847 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hsrtq" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.793947 4847 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hsrtq" event={"ID":"2f9f5bc0-dba0-4c11-a612-5ab327ed22b2","Type":"ContainerDied","Data":"42ce366b574660e75e678ae4dc693e0a27e218d7bf16b0170a7d1fdded6362ba"} Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.793968 4847 scope.go:117] "RemoveContainer" containerID="c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.817726 4847 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.822677 4847 scope.go:117] "RemoveContainer" containerID="1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.826772 4847 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hsrtq"] Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.842186 4847 scope.go:117] "RemoveContainer" containerID="d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.881448 4847 scope.go:117] "RemoveContainer" containerID="c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9" Dec 10 15:38:28 crc kubenswrapper[4847]: E1210 15:38:28.882056 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9\": container with ID starting with c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9 not found: ID does not exist" containerID="c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.882101 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9"} err="failed to get container status \"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9\": rpc error: code = NotFound desc = could not find container \"c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9\": container with ID starting with c0029f935c1cb070031eed1ddd07e3b7c938d696b48d8d42769129e326feb5f9 not found: ID does not exist" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.882128 4847 scope.go:117] "RemoveContainer" containerID="1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f" Dec 10 15:38:28 crc kubenswrapper[4847]: E1210 15:38:28.882563 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f\": container with ID starting with 1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f not found: ID does not exist" containerID="1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.882594 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f"} err="failed to get container status \"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f\": rpc error: code = NotFound desc = could not find container \"1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f\": container with ID starting with 1e7f9c5267081192f1958ab8b26235705b79ca308ddb6f2d049d39199530173f not found: ID does not exist" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.882613 4847 scope.go:117] "RemoveContainer" containerID="d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b" Dec 10 15:38:28 crc kubenswrapper[4847]: E1210 15:38:28.882944 4847 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b\": container with ID starting with d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b not found: ID does not exist" containerID="d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b" Dec 10 15:38:28 crc kubenswrapper[4847]: I1210 15:38:28.882976 4847 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b"} err="failed to get container status \"d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b\": rpc error: code = NotFound desc = could not find container \"d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b\": container with ID starting with d8b67c4f5ebe2413262f4bd76df30fe644c55fc533363e888b445340d406d60b not found: ID does not exist" Dec 10 15:38:30 crc kubenswrapper[4847]: I1210 15:38:30.774125 4847 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f9f5bc0-dba0-4c11-a612-5ab327ed22b2" path="/var/lib/kubelet/pods/2f9f5bc0-dba0-4c11-a612-5ab327ed22b2/volumes" Dec 10 15:38:33 crc kubenswrapper[4847]: I1210 15:38:33.759984 4847 scope.go:117] "RemoveContainer" containerID="ba7bc1ec0f72c015dccdbb30ee4ca6a97e14188b4dd0fc1cf031fdd558df430b" Dec 10 15:38:33 crc kubenswrapper[4847]: E1210 15:38:33.760838 4847 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gvdwq_openshift-machine-config-operator(8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324)\"" pod="openshift-machine-config-operator/machine-config-daemon-gvdwq" podUID="8a8a5b2d-2b36-4f87-bec6-8ecb7fe32324"